{ "best_metric": null, "best_model_checkpoint": null, "epoch": 1.0, "eval_steps": 500, "global_step": 9253, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.00010807305738679347, "grad_norm": 0.4351649880409241, "learning_rate": 8e-05, "loss": 1.8983, "step": 1 }, { "epoch": 0.00021614611477358694, "grad_norm": 0.5632684230804443, "learning_rate": 8e-05, "loss": 2.0631, "step": 2 }, { "epoch": 0.0003242191721603804, "grad_norm": 0.4193551540374756, "learning_rate": 8e-05, "loss": 1.9084, "step": 3 }, { "epoch": 0.0004322922295471739, "grad_norm": 0.43743789196014404, "learning_rate": 8e-05, "loss": 2.0345, "step": 4 }, { "epoch": 0.0005403652869339674, "grad_norm": 0.35983482003211975, "learning_rate": 8e-05, "loss": 1.9538, "step": 5 }, { "epoch": 0.0006484383443207608, "grad_norm": 0.37199866771698, "learning_rate": 8e-05, "loss": 2.1294, "step": 6 }, { "epoch": 0.0007565114017075543, "grad_norm": 0.35996878147125244, "learning_rate": 8e-05, "loss": 1.8098, "step": 7 }, { "epoch": 0.0008645844590943478, "grad_norm": 0.40022915601730347, "learning_rate": 8e-05, "loss": 2.0355, "step": 8 }, { "epoch": 0.0009726575164811412, "grad_norm": 0.4122651517391205, "learning_rate": 8e-05, "loss": 1.9823, "step": 9 }, { "epoch": 0.0010807305738679347, "grad_norm": 0.43956854939460754, "learning_rate": 8e-05, "loss": 1.858, "step": 10 }, { "epoch": 0.0011888036312547282, "grad_norm": 0.398956835269928, "learning_rate": 8e-05, "loss": 1.9997, "step": 11 }, { "epoch": 0.0012968766886415216, "grad_norm": 0.34920763969421387, "learning_rate": 8e-05, "loss": 1.7847, "step": 12 }, { "epoch": 0.001404949746028315, "grad_norm": 0.35869136452674866, "learning_rate": 8e-05, "loss": 1.8985, "step": 13 }, { "epoch": 0.0015130228034151087, "grad_norm": 0.32896265387535095, "learning_rate": 8e-05, "loss": 1.7711, "step": 14 }, { "epoch": 0.0016210958608019021, "grad_norm": 0.36838844418525696, "learning_rate": 8e-05, "loss": 1.8666, "step": 15 }, { "epoch": 0.0017291689181886955, "grad_norm": 0.3589898943901062, "learning_rate": 8e-05, "loss": 1.8331, "step": 16 }, { "epoch": 0.001837241975575489, "grad_norm": 0.3534690737724304, "learning_rate": 8e-05, "loss": 1.8592, "step": 17 }, { "epoch": 0.0019453150329622824, "grad_norm": 0.33605045080184937, "learning_rate": 8e-05, "loss": 1.8403, "step": 18 }, { "epoch": 0.002053388090349076, "grad_norm": 0.38296014070510864, "learning_rate": 8e-05, "loss": 2.063, "step": 19 }, { "epoch": 0.0021614611477358695, "grad_norm": 0.356327623128891, "learning_rate": 8e-05, "loss": 1.8731, "step": 20 }, { "epoch": 0.002269534205122663, "grad_norm": 0.3825637102127075, "learning_rate": 8e-05, "loss": 1.8076, "step": 21 }, { "epoch": 0.0023776072625094563, "grad_norm": 0.3552597463130951, "learning_rate": 8e-05, "loss": 1.7152, "step": 22 }, { "epoch": 0.0024856803198962498, "grad_norm": 0.3945937156677246, "learning_rate": 8e-05, "loss": 2.0094, "step": 23 }, { "epoch": 0.002593753377283043, "grad_norm": 0.3497529923915863, "learning_rate": 8e-05, "loss": 1.8048, "step": 24 }, { "epoch": 0.0027018264346698366, "grad_norm": 0.35496705770492554, "learning_rate": 8e-05, "loss": 1.9212, "step": 25 }, { "epoch": 0.00280989949205663, "grad_norm": 0.40153196454048157, "learning_rate": 8e-05, "loss": 2.0488, "step": 26 }, { "epoch": 0.002917972549443424, "grad_norm": 0.3893981873989105, "learning_rate": 8e-05, "loss": 1.8708, "step": 27 }, { "epoch": 0.0030260456068302174, "grad_norm": 0.3695599138736725, "learning_rate": 8e-05, "loss": 2.0791, "step": 28 }, { "epoch": 0.003134118664217011, "grad_norm": 0.3641146421432495, "learning_rate": 8e-05, "loss": 1.9072, "step": 29 }, { "epoch": 0.0032421917216038042, "grad_norm": 0.3596847355365753, "learning_rate": 8e-05, "loss": 1.8498, "step": 30 }, { "epoch": 0.0033502647789905977, "grad_norm": 0.3638761341571808, "learning_rate": 8e-05, "loss": 1.8096, "step": 31 }, { "epoch": 0.003458337836377391, "grad_norm": 0.39277321100234985, "learning_rate": 8e-05, "loss": 1.8466, "step": 32 }, { "epoch": 0.0035664108937641845, "grad_norm": 0.3584340214729309, "learning_rate": 8e-05, "loss": 1.9072, "step": 33 }, { "epoch": 0.003674483951150978, "grad_norm": 0.38852962851524353, "learning_rate": 8e-05, "loss": 1.9932, "step": 34 }, { "epoch": 0.0037825570085377714, "grad_norm": 0.39563801884651184, "learning_rate": 8e-05, "loss": 1.8998, "step": 35 }, { "epoch": 0.003890630065924565, "grad_norm": 0.3966459035873413, "learning_rate": 8e-05, "loss": 1.941, "step": 36 }, { "epoch": 0.003998703123311359, "grad_norm": 0.3564274311065674, "learning_rate": 8e-05, "loss": 1.8725, "step": 37 }, { "epoch": 0.004106776180698152, "grad_norm": 0.34590858221054077, "learning_rate": 8e-05, "loss": 1.8007, "step": 38 }, { "epoch": 0.0042148492380849455, "grad_norm": 0.3475266993045807, "learning_rate": 8e-05, "loss": 1.8199, "step": 39 }, { "epoch": 0.004322922295471739, "grad_norm": 0.36428943276405334, "learning_rate": 8e-05, "loss": 1.8987, "step": 40 }, { "epoch": 0.004430995352858532, "grad_norm": 0.36737462878227234, "learning_rate": 8e-05, "loss": 1.9232, "step": 41 }, { "epoch": 0.004539068410245326, "grad_norm": 0.41655054688453674, "learning_rate": 8e-05, "loss": 1.9765, "step": 42 }, { "epoch": 0.004647141467632119, "grad_norm": 0.37395918369293213, "learning_rate": 8e-05, "loss": 1.9251, "step": 43 }, { "epoch": 0.004755214525018913, "grad_norm": 0.36078399419784546, "learning_rate": 8e-05, "loss": 1.9252, "step": 44 }, { "epoch": 0.004863287582405706, "grad_norm": 0.37375158071517944, "learning_rate": 8e-05, "loss": 1.9898, "step": 45 }, { "epoch": 0.0049713606397924996, "grad_norm": 0.35973864793777466, "learning_rate": 8e-05, "loss": 1.8468, "step": 46 }, { "epoch": 0.005079433697179293, "grad_norm": 0.33282211422920227, "learning_rate": 8e-05, "loss": 1.8168, "step": 47 }, { "epoch": 0.005187506754566086, "grad_norm": 0.3648369312286377, "learning_rate": 8e-05, "loss": 1.8562, "step": 48 }, { "epoch": 0.00529557981195288, "grad_norm": 0.3680543005466461, "learning_rate": 8e-05, "loss": 1.7302, "step": 49 }, { "epoch": 0.005403652869339673, "grad_norm": 0.3996305465698242, "learning_rate": 8e-05, "loss": 2.0178, "step": 50 }, { "epoch": 0.005511725926726467, "grad_norm": 0.34245601296424866, "learning_rate": 8e-05, "loss": 1.7272, "step": 51 }, { "epoch": 0.00561979898411326, "grad_norm": 0.4385335147380829, "learning_rate": 8e-05, "loss": 2.1892, "step": 52 }, { "epoch": 0.005727872041500054, "grad_norm": 0.37781935930252075, "learning_rate": 8e-05, "loss": 1.8742, "step": 53 }, { "epoch": 0.005835945098886848, "grad_norm": 0.35104045271873474, "learning_rate": 8e-05, "loss": 1.8235, "step": 54 }, { "epoch": 0.005944018156273641, "grad_norm": 0.3527902066707611, "learning_rate": 8e-05, "loss": 1.8316, "step": 55 }, { "epoch": 0.006052091213660435, "grad_norm": 0.3233564794063568, "learning_rate": 8e-05, "loss": 1.8116, "step": 56 }, { "epoch": 0.006160164271047228, "grad_norm": 0.36856353282928467, "learning_rate": 8e-05, "loss": 1.7645, "step": 57 }, { "epoch": 0.006268237328434022, "grad_norm": 0.4024446904659271, "learning_rate": 8e-05, "loss": 1.939, "step": 58 }, { "epoch": 0.006376310385820815, "grad_norm": 0.3551691174507141, "learning_rate": 8e-05, "loss": 1.8413, "step": 59 }, { "epoch": 0.0064843834432076084, "grad_norm": 0.44394415616989136, "learning_rate": 8e-05, "loss": 2.2443, "step": 60 }, { "epoch": 0.006592456500594402, "grad_norm": 0.3416070342063904, "learning_rate": 8e-05, "loss": 1.5767, "step": 61 }, { "epoch": 0.006700529557981195, "grad_norm": 0.3482869565486908, "learning_rate": 8e-05, "loss": 1.8669, "step": 62 }, { "epoch": 0.006808602615367989, "grad_norm": 0.3702720105648041, "learning_rate": 8e-05, "loss": 1.7849, "step": 63 }, { "epoch": 0.006916675672754782, "grad_norm": 0.36599811911582947, "learning_rate": 8e-05, "loss": 2.0098, "step": 64 }, { "epoch": 0.007024748730141576, "grad_norm": 0.3971049189567566, "learning_rate": 8e-05, "loss": 2.0508, "step": 65 }, { "epoch": 0.007132821787528369, "grad_norm": 0.35264477133750916, "learning_rate": 8e-05, "loss": 1.7954, "step": 66 }, { "epoch": 0.0072408948449151625, "grad_norm": 0.3791993260383606, "learning_rate": 8e-05, "loss": 1.9896, "step": 67 }, { "epoch": 0.007348967902301956, "grad_norm": 0.37191087007522583, "learning_rate": 8e-05, "loss": 1.9054, "step": 68 }, { "epoch": 0.007457040959688749, "grad_norm": 0.4020589590072632, "learning_rate": 8e-05, "loss": 1.7792, "step": 69 }, { "epoch": 0.007565114017075543, "grad_norm": 0.34911972284317017, "learning_rate": 8e-05, "loss": 1.8529, "step": 70 }, { "epoch": 0.007673187074462336, "grad_norm": 0.37512168288230896, "learning_rate": 8e-05, "loss": 1.9126, "step": 71 }, { "epoch": 0.00778126013184913, "grad_norm": 0.3415440618991852, "learning_rate": 8e-05, "loss": 1.8138, "step": 72 }, { "epoch": 0.007889333189235924, "grad_norm": 0.3596620559692383, "learning_rate": 8e-05, "loss": 1.959, "step": 73 }, { "epoch": 0.007997406246622717, "grad_norm": 0.3739068806171417, "learning_rate": 8e-05, "loss": 1.8026, "step": 74 }, { "epoch": 0.00810547930400951, "grad_norm": 0.33997228741645813, "learning_rate": 8e-05, "loss": 1.8809, "step": 75 }, { "epoch": 0.008213552361396304, "grad_norm": 0.3936220407485962, "learning_rate": 8e-05, "loss": 1.9793, "step": 76 }, { "epoch": 0.008321625418783098, "grad_norm": 0.3442775011062622, "learning_rate": 8e-05, "loss": 1.8598, "step": 77 }, { "epoch": 0.008429698476169891, "grad_norm": 0.34543630480766296, "learning_rate": 8e-05, "loss": 1.8743, "step": 78 }, { "epoch": 0.008537771533556685, "grad_norm": 0.3811757266521454, "learning_rate": 8e-05, "loss": 1.9644, "step": 79 }, { "epoch": 0.008645844590943478, "grad_norm": 0.3636032044887543, "learning_rate": 8e-05, "loss": 1.9381, "step": 80 }, { "epoch": 0.008753917648330271, "grad_norm": 0.3840232193470001, "learning_rate": 8e-05, "loss": 2.0628, "step": 81 }, { "epoch": 0.008861990705717065, "grad_norm": 0.3367981016635895, "learning_rate": 8e-05, "loss": 1.782, "step": 82 }, { "epoch": 0.008970063763103858, "grad_norm": 0.33157339692115784, "learning_rate": 8e-05, "loss": 1.8371, "step": 83 }, { "epoch": 0.009078136820490652, "grad_norm": 0.3375696837902069, "learning_rate": 8e-05, "loss": 1.8922, "step": 84 }, { "epoch": 0.009186209877877445, "grad_norm": 0.3356796205043793, "learning_rate": 8e-05, "loss": 1.6773, "step": 85 }, { "epoch": 0.009294282935264239, "grad_norm": 0.3588196635246277, "learning_rate": 8e-05, "loss": 1.7572, "step": 86 }, { "epoch": 0.009402355992651032, "grad_norm": 0.3561270534992218, "learning_rate": 8e-05, "loss": 1.9395, "step": 87 }, { "epoch": 0.009510429050037825, "grad_norm": 0.3281303942203522, "learning_rate": 8e-05, "loss": 1.574, "step": 88 }, { "epoch": 0.009618502107424619, "grad_norm": 0.34098535776138306, "learning_rate": 8e-05, "loss": 1.7778, "step": 89 }, { "epoch": 0.009726575164811412, "grad_norm": 0.3468838334083557, "learning_rate": 8e-05, "loss": 1.7544, "step": 90 }, { "epoch": 0.009834648222198206, "grad_norm": 0.3873237073421478, "learning_rate": 8e-05, "loss": 1.8939, "step": 91 }, { "epoch": 0.009942721279584999, "grad_norm": 0.34803327918052673, "learning_rate": 8e-05, "loss": 1.923, "step": 92 }, { "epoch": 0.010050794336971793, "grad_norm": 0.3591421842575073, "learning_rate": 8e-05, "loss": 1.8173, "step": 93 }, { "epoch": 0.010158867394358586, "grad_norm": 0.3685607314109802, "learning_rate": 8e-05, "loss": 1.8363, "step": 94 }, { "epoch": 0.01026694045174538, "grad_norm": 0.4085378050804138, "learning_rate": 8e-05, "loss": 2.0675, "step": 95 }, { "epoch": 0.010375013509132173, "grad_norm": 0.33424437046051025, "learning_rate": 8e-05, "loss": 1.8288, "step": 96 }, { "epoch": 0.010483086566518966, "grad_norm": 0.3550828993320465, "learning_rate": 8e-05, "loss": 1.7532, "step": 97 }, { "epoch": 0.01059115962390576, "grad_norm": 0.3696897327899933, "learning_rate": 8e-05, "loss": 1.9669, "step": 98 }, { "epoch": 0.010699232681292553, "grad_norm": 0.33994463086128235, "learning_rate": 8e-05, "loss": 1.6709, "step": 99 }, { "epoch": 0.010807305738679347, "grad_norm": 0.36720365285873413, "learning_rate": 8e-05, "loss": 2.028, "step": 100 }, { "epoch": 0.01091537879606614, "grad_norm": 0.3576091527938843, "learning_rate": 8e-05, "loss": 1.6761, "step": 101 }, { "epoch": 0.011023451853452933, "grad_norm": 0.3334646224975586, "learning_rate": 8e-05, "loss": 1.7336, "step": 102 }, { "epoch": 0.011131524910839727, "grad_norm": 0.39445286989212036, "learning_rate": 8e-05, "loss": 1.9391, "step": 103 }, { "epoch": 0.01123959796822652, "grad_norm": 0.34280478954315186, "learning_rate": 8e-05, "loss": 1.933, "step": 104 }, { "epoch": 0.011347671025613315, "grad_norm": 0.36647966504096985, "learning_rate": 8e-05, "loss": 1.9672, "step": 105 }, { "epoch": 0.011455744083000109, "grad_norm": 0.4370229244232178, "learning_rate": 8e-05, "loss": 1.7903, "step": 106 }, { "epoch": 0.011563817140386902, "grad_norm": 0.38026702404022217, "learning_rate": 8e-05, "loss": 1.9277, "step": 107 }, { "epoch": 0.011671890197773696, "grad_norm": 0.3587406277656555, "learning_rate": 8e-05, "loss": 1.9038, "step": 108 }, { "epoch": 0.01177996325516049, "grad_norm": 0.3514784872531891, "learning_rate": 8e-05, "loss": 1.7971, "step": 109 }, { "epoch": 0.011888036312547283, "grad_norm": 0.3730924427509308, "learning_rate": 8e-05, "loss": 1.6484, "step": 110 }, { "epoch": 0.011996109369934076, "grad_norm": 0.3412942886352539, "learning_rate": 8e-05, "loss": 1.8384, "step": 111 }, { "epoch": 0.01210418242732087, "grad_norm": 0.36606597900390625, "learning_rate": 8e-05, "loss": 1.9919, "step": 112 }, { "epoch": 0.012212255484707663, "grad_norm": 0.3667832911014557, "learning_rate": 8e-05, "loss": 1.7714, "step": 113 }, { "epoch": 0.012320328542094456, "grad_norm": 0.3357333540916443, "learning_rate": 8e-05, "loss": 1.7315, "step": 114 }, { "epoch": 0.01242840159948125, "grad_norm": 0.3732669949531555, "learning_rate": 8e-05, "loss": 1.941, "step": 115 }, { "epoch": 0.012536474656868043, "grad_norm": 0.48029825091362, "learning_rate": 8e-05, "loss": 1.7861, "step": 116 }, { "epoch": 0.012644547714254837, "grad_norm": 0.3361777663230896, "learning_rate": 8e-05, "loss": 1.7655, "step": 117 }, { "epoch": 0.01275262077164163, "grad_norm": 0.3681136667728424, "learning_rate": 8e-05, "loss": 1.8102, "step": 118 }, { "epoch": 0.012860693829028423, "grad_norm": 0.4084520637989044, "learning_rate": 8e-05, "loss": 1.9245, "step": 119 }, { "epoch": 0.012968766886415217, "grad_norm": 0.34002935886383057, "learning_rate": 8e-05, "loss": 1.8387, "step": 120 }, { "epoch": 0.01307683994380201, "grad_norm": 0.35462895035743713, "learning_rate": 8e-05, "loss": 1.912, "step": 121 }, { "epoch": 0.013184913001188804, "grad_norm": 0.3921608328819275, "learning_rate": 8e-05, "loss": 1.8677, "step": 122 }, { "epoch": 0.013292986058575597, "grad_norm": 0.3691435754299164, "learning_rate": 8e-05, "loss": 1.8201, "step": 123 }, { "epoch": 0.01340105911596239, "grad_norm": 0.3534267544746399, "learning_rate": 8e-05, "loss": 1.8456, "step": 124 }, { "epoch": 0.013509132173349184, "grad_norm": 0.35075145959854126, "learning_rate": 8e-05, "loss": 1.8054, "step": 125 }, { "epoch": 0.013617205230735977, "grad_norm": 0.37121960520744324, "learning_rate": 8e-05, "loss": 1.8728, "step": 126 }, { "epoch": 0.013725278288122771, "grad_norm": 0.4479074478149414, "learning_rate": 8e-05, "loss": 1.972, "step": 127 }, { "epoch": 0.013833351345509564, "grad_norm": 0.3476560413837433, "learning_rate": 8e-05, "loss": 1.8225, "step": 128 }, { "epoch": 0.013941424402896358, "grad_norm": 0.3355270028114319, "learning_rate": 8e-05, "loss": 1.8436, "step": 129 }, { "epoch": 0.014049497460283151, "grad_norm": 0.42383939027786255, "learning_rate": 8e-05, "loss": 1.9533, "step": 130 }, { "epoch": 0.014157570517669945, "grad_norm": 0.3658997416496277, "learning_rate": 8e-05, "loss": 1.943, "step": 131 }, { "epoch": 0.014265643575056738, "grad_norm": 0.38565921783447266, "learning_rate": 8e-05, "loss": 1.9349, "step": 132 }, { "epoch": 0.014373716632443531, "grad_norm": 0.390728622674942, "learning_rate": 8e-05, "loss": 1.882, "step": 133 }, { "epoch": 0.014481789689830325, "grad_norm": 0.37281593680381775, "learning_rate": 8e-05, "loss": 1.7955, "step": 134 }, { "epoch": 0.014589862747217118, "grad_norm": 0.42595094442367554, "learning_rate": 8e-05, "loss": 1.9489, "step": 135 }, { "epoch": 0.014697935804603912, "grad_norm": 0.3504122495651245, "learning_rate": 8e-05, "loss": 1.8774, "step": 136 }, { "epoch": 0.014806008861990705, "grad_norm": 0.4228856861591339, "learning_rate": 8e-05, "loss": 2.0514, "step": 137 }, { "epoch": 0.014914081919377499, "grad_norm": 0.3680881857872009, "learning_rate": 8e-05, "loss": 1.86, "step": 138 }, { "epoch": 0.015022154976764292, "grad_norm": 0.344536155462265, "learning_rate": 8e-05, "loss": 1.8319, "step": 139 }, { "epoch": 0.015130228034151086, "grad_norm": 0.3578692376613617, "learning_rate": 8e-05, "loss": 1.8546, "step": 140 }, { "epoch": 0.015238301091537879, "grad_norm": 0.42693620920181274, "learning_rate": 8e-05, "loss": 1.763, "step": 141 }, { "epoch": 0.015346374148924672, "grad_norm": 0.4166564345359802, "learning_rate": 8e-05, "loss": 1.9141, "step": 142 }, { "epoch": 0.015454447206311466, "grad_norm": 0.38682886958122253, "learning_rate": 8e-05, "loss": 1.8005, "step": 143 }, { "epoch": 0.01556252026369826, "grad_norm": 0.33764222264289856, "learning_rate": 8e-05, "loss": 1.7668, "step": 144 }, { "epoch": 0.015670593321085054, "grad_norm": 0.3671913146972656, "learning_rate": 8e-05, "loss": 1.8183, "step": 145 }, { "epoch": 0.015778666378471848, "grad_norm": 0.39653363823890686, "learning_rate": 8e-05, "loss": 1.9931, "step": 146 }, { "epoch": 0.01588673943585864, "grad_norm": 0.39502009749412537, "learning_rate": 8e-05, "loss": 1.8918, "step": 147 }, { "epoch": 0.015994812493245435, "grad_norm": 0.35358163714408875, "learning_rate": 8e-05, "loss": 1.9298, "step": 148 }, { "epoch": 0.016102885550632228, "grad_norm": 0.35190215706825256, "learning_rate": 8e-05, "loss": 1.7793, "step": 149 }, { "epoch": 0.01621095860801902, "grad_norm": 0.3853912055492401, "learning_rate": 8e-05, "loss": 1.8762, "step": 150 }, { "epoch": 0.016319031665405815, "grad_norm": 0.3836837410926819, "learning_rate": 8e-05, "loss": 1.8991, "step": 151 }, { "epoch": 0.01642710472279261, "grad_norm": 0.3867194950580597, "learning_rate": 8e-05, "loss": 2.0555, "step": 152 }, { "epoch": 0.016535177780179402, "grad_norm": 0.3466987907886505, "learning_rate": 8e-05, "loss": 1.7332, "step": 153 }, { "epoch": 0.016643250837566195, "grad_norm": 0.4112001359462738, "learning_rate": 8e-05, "loss": 1.7648, "step": 154 }, { "epoch": 0.01675132389495299, "grad_norm": 0.36368218064308167, "learning_rate": 8e-05, "loss": 1.8836, "step": 155 }, { "epoch": 0.016859396952339782, "grad_norm": 0.343652606010437, "learning_rate": 8e-05, "loss": 1.763, "step": 156 }, { "epoch": 0.016967470009726576, "grad_norm": 0.3557482957839966, "learning_rate": 8e-05, "loss": 1.8708, "step": 157 }, { "epoch": 0.01707554306711337, "grad_norm": 0.3625258207321167, "learning_rate": 8e-05, "loss": 1.86, "step": 158 }, { "epoch": 0.017183616124500162, "grad_norm": 0.4034252166748047, "learning_rate": 8e-05, "loss": 2.0207, "step": 159 }, { "epoch": 0.017291689181886956, "grad_norm": 0.3633774220943451, "learning_rate": 8e-05, "loss": 1.8668, "step": 160 }, { "epoch": 0.01739976223927375, "grad_norm": 0.326499342918396, "learning_rate": 8e-05, "loss": 1.75, "step": 161 }, { "epoch": 0.017507835296660543, "grad_norm": 0.33330976963043213, "learning_rate": 8e-05, "loss": 1.773, "step": 162 }, { "epoch": 0.017615908354047336, "grad_norm": 0.37760624289512634, "learning_rate": 8e-05, "loss": 1.9396, "step": 163 }, { "epoch": 0.01772398141143413, "grad_norm": 0.32992613315582275, "learning_rate": 8e-05, "loss": 1.7772, "step": 164 }, { "epoch": 0.017832054468820923, "grad_norm": 0.35324472188949585, "learning_rate": 8e-05, "loss": 1.9957, "step": 165 }, { "epoch": 0.017940127526207716, "grad_norm": 0.33778947591781616, "learning_rate": 8e-05, "loss": 1.7158, "step": 166 }, { "epoch": 0.01804820058359451, "grad_norm": 0.35893818736076355, "learning_rate": 8e-05, "loss": 1.9277, "step": 167 }, { "epoch": 0.018156273640981303, "grad_norm": 0.37994202971458435, "learning_rate": 8e-05, "loss": 2.0575, "step": 168 }, { "epoch": 0.018264346698368097, "grad_norm": 0.35307133197784424, "learning_rate": 8e-05, "loss": 1.7618, "step": 169 }, { "epoch": 0.01837241975575489, "grad_norm": 0.3657943904399872, "learning_rate": 8e-05, "loss": 1.7863, "step": 170 }, { "epoch": 0.018480492813141684, "grad_norm": 0.3480663597583771, "learning_rate": 8e-05, "loss": 1.7693, "step": 171 }, { "epoch": 0.018588565870528477, "grad_norm": 0.5380162596702576, "learning_rate": 8e-05, "loss": 2.1205, "step": 172 }, { "epoch": 0.01869663892791527, "grad_norm": 0.3727838695049286, "learning_rate": 8e-05, "loss": 1.8845, "step": 173 }, { "epoch": 0.018804711985302064, "grad_norm": 0.3410198390483856, "learning_rate": 8e-05, "loss": 1.7974, "step": 174 }, { "epoch": 0.018912785042688857, "grad_norm": 0.33915817737579346, "learning_rate": 8e-05, "loss": 1.7681, "step": 175 }, { "epoch": 0.01902085810007565, "grad_norm": 0.34816646575927734, "learning_rate": 8e-05, "loss": 1.8082, "step": 176 }, { "epoch": 0.019128931157462444, "grad_norm": 0.3584868311882019, "learning_rate": 8e-05, "loss": 1.8774, "step": 177 }, { "epoch": 0.019237004214849238, "grad_norm": 0.3616390526294708, "learning_rate": 8e-05, "loss": 1.6646, "step": 178 }, { "epoch": 0.01934507727223603, "grad_norm": 0.3825621008872986, "learning_rate": 8e-05, "loss": 1.8342, "step": 179 }, { "epoch": 0.019453150329622824, "grad_norm": 0.37360844016075134, "learning_rate": 8e-05, "loss": 1.9066, "step": 180 }, { "epoch": 0.019561223387009618, "grad_norm": 0.3455824553966522, "learning_rate": 8e-05, "loss": 1.7934, "step": 181 }, { "epoch": 0.01966929644439641, "grad_norm": 0.35564401745796204, "learning_rate": 8e-05, "loss": 1.8435, "step": 182 }, { "epoch": 0.019777369501783205, "grad_norm": 0.3567356467247009, "learning_rate": 8e-05, "loss": 1.854, "step": 183 }, { "epoch": 0.019885442559169998, "grad_norm": 0.3364497125148773, "learning_rate": 8e-05, "loss": 1.7849, "step": 184 }, { "epoch": 0.01999351561655679, "grad_norm": 0.3673102855682373, "learning_rate": 8e-05, "loss": 2.044, "step": 185 }, { "epoch": 0.020101588673943585, "grad_norm": 0.3575238287448883, "learning_rate": 8e-05, "loss": 1.862, "step": 186 }, { "epoch": 0.02020966173133038, "grad_norm": 0.3294106423854828, "learning_rate": 8e-05, "loss": 1.7408, "step": 187 }, { "epoch": 0.020317734788717172, "grad_norm": 0.36280909180641174, "learning_rate": 8e-05, "loss": 1.9266, "step": 188 }, { "epoch": 0.020425807846103965, "grad_norm": 0.3469875454902649, "learning_rate": 8e-05, "loss": 1.9141, "step": 189 }, { "epoch": 0.02053388090349076, "grad_norm": 0.35601651668548584, "learning_rate": 8e-05, "loss": 1.9005, "step": 190 }, { "epoch": 0.020641953960877552, "grad_norm": 0.3853970766067505, "learning_rate": 8e-05, "loss": 1.9129, "step": 191 }, { "epoch": 0.020750027018264346, "grad_norm": 0.37526872754096985, "learning_rate": 8e-05, "loss": 1.9607, "step": 192 }, { "epoch": 0.02085810007565114, "grad_norm": 0.36681321263313293, "learning_rate": 8e-05, "loss": 1.9772, "step": 193 }, { "epoch": 0.020966173133037933, "grad_norm": 0.3470933735370636, "learning_rate": 8e-05, "loss": 1.9596, "step": 194 }, { "epoch": 0.021074246190424726, "grad_norm": 0.3369242250919342, "learning_rate": 8e-05, "loss": 1.777, "step": 195 }, { "epoch": 0.02118231924781152, "grad_norm": 0.3203667402267456, "learning_rate": 8e-05, "loss": 1.6908, "step": 196 }, { "epoch": 0.021290392305198313, "grad_norm": 0.3353847563266754, "learning_rate": 8e-05, "loss": 1.8597, "step": 197 }, { "epoch": 0.021398465362585106, "grad_norm": 0.3403714597225189, "learning_rate": 8e-05, "loss": 1.9415, "step": 198 }, { "epoch": 0.0215065384199719, "grad_norm": 0.33768534660339355, "learning_rate": 8e-05, "loss": 1.8627, "step": 199 }, { "epoch": 0.021614611477358693, "grad_norm": 0.3423212766647339, "learning_rate": 8e-05, "loss": 1.8255, "step": 200 }, { "epoch": 0.021722684534745487, "grad_norm": 0.33419105410575867, "learning_rate": 8e-05, "loss": 1.8485, "step": 201 }, { "epoch": 0.02183075759213228, "grad_norm": 0.31438854336738586, "learning_rate": 8e-05, "loss": 1.7767, "step": 202 }, { "epoch": 0.021938830649519073, "grad_norm": 0.3373056650161743, "learning_rate": 8e-05, "loss": 1.7299, "step": 203 }, { "epoch": 0.022046903706905867, "grad_norm": 0.3650558292865753, "learning_rate": 8e-05, "loss": 1.8846, "step": 204 }, { "epoch": 0.02215497676429266, "grad_norm": 0.3386206030845642, "learning_rate": 8e-05, "loss": 1.7638, "step": 205 }, { "epoch": 0.022263049821679454, "grad_norm": 0.35893094539642334, "learning_rate": 8e-05, "loss": 1.8466, "step": 206 }, { "epoch": 0.022371122879066247, "grad_norm": 0.33917978405952454, "learning_rate": 8e-05, "loss": 1.7422, "step": 207 }, { "epoch": 0.02247919593645304, "grad_norm": 0.378278911113739, "learning_rate": 8e-05, "loss": 1.9303, "step": 208 }, { "epoch": 0.022587268993839837, "grad_norm": 0.3625754415988922, "learning_rate": 8e-05, "loss": 1.814, "step": 209 }, { "epoch": 0.02269534205122663, "grad_norm": 0.4043956696987152, "learning_rate": 8e-05, "loss": 2.0295, "step": 210 }, { "epoch": 0.022803415108613424, "grad_norm": 0.34033194184303284, "learning_rate": 8e-05, "loss": 1.7618, "step": 211 }, { "epoch": 0.022911488166000218, "grad_norm": 0.35087522864341736, "learning_rate": 8e-05, "loss": 1.9093, "step": 212 }, { "epoch": 0.02301956122338701, "grad_norm": 0.3271329998970032, "learning_rate": 8e-05, "loss": 1.6803, "step": 213 }, { "epoch": 0.023127634280773805, "grad_norm": 0.3723451793193817, "learning_rate": 8e-05, "loss": 1.9597, "step": 214 }, { "epoch": 0.023235707338160598, "grad_norm": 0.36175742745399475, "learning_rate": 8e-05, "loss": 1.6835, "step": 215 }, { "epoch": 0.02334378039554739, "grad_norm": 0.3611111044883728, "learning_rate": 8e-05, "loss": 1.7868, "step": 216 }, { "epoch": 0.023451853452934185, "grad_norm": 0.3544881343841553, "learning_rate": 8e-05, "loss": 1.802, "step": 217 }, { "epoch": 0.02355992651032098, "grad_norm": 0.3522435426712036, "learning_rate": 8e-05, "loss": 1.8659, "step": 218 }, { "epoch": 0.023667999567707772, "grad_norm": 0.3727189898490906, "learning_rate": 8e-05, "loss": 1.927, "step": 219 }, { "epoch": 0.023776072625094565, "grad_norm": 0.3585643768310547, "learning_rate": 8e-05, "loss": 1.7955, "step": 220 }, { "epoch": 0.02388414568248136, "grad_norm": 0.3653928339481354, "learning_rate": 8e-05, "loss": 1.8583, "step": 221 }, { "epoch": 0.023992218739868152, "grad_norm": 0.3753211200237274, "learning_rate": 8e-05, "loss": 1.6564, "step": 222 }, { "epoch": 0.024100291797254945, "grad_norm": 0.3349682688713074, "learning_rate": 8e-05, "loss": 1.7643, "step": 223 }, { "epoch": 0.02420836485464174, "grad_norm": 0.3413155674934387, "learning_rate": 8e-05, "loss": 1.7041, "step": 224 }, { "epoch": 0.024316437912028532, "grad_norm": 0.33893683552742004, "learning_rate": 8e-05, "loss": 1.6973, "step": 225 }, { "epoch": 0.024424510969415326, "grad_norm": 0.3363328278064728, "learning_rate": 8e-05, "loss": 1.8412, "step": 226 }, { "epoch": 0.02453258402680212, "grad_norm": 0.3394852876663208, "learning_rate": 8e-05, "loss": 1.7423, "step": 227 }, { "epoch": 0.024640657084188913, "grad_norm": 0.3761345446109772, "learning_rate": 8e-05, "loss": 1.9233, "step": 228 }, { "epoch": 0.024748730141575706, "grad_norm": 0.3359405994415283, "learning_rate": 8e-05, "loss": 1.7564, "step": 229 }, { "epoch": 0.0248568031989625, "grad_norm": 0.35076820850372314, "learning_rate": 8e-05, "loss": 1.7561, "step": 230 }, { "epoch": 0.024964876256349293, "grad_norm": 0.3408905267715454, "learning_rate": 8e-05, "loss": 1.7931, "step": 231 }, { "epoch": 0.025072949313736086, "grad_norm": 0.3573218584060669, "learning_rate": 8e-05, "loss": 1.8407, "step": 232 }, { "epoch": 0.02518102237112288, "grad_norm": 0.31585562229156494, "learning_rate": 8e-05, "loss": 1.713, "step": 233 }, { "epoch": 0.025289095428509673, "grad_norm": 0.3589065372943878, "learning_rate": 8e-05, "loss": 1.9227, "step": 234 }, { "epoch": 0.025397168485896467, "grad_norm": 0.3374887704849243, "learning_rate": 8e-05, "loss": 1.7683, "step": 235 }, { "epoch": 0.02550524154328326, "grad_norm": 0.35544005036354065, "learning_rate": 8e-05, "loss": 1.8151, "step": 236 }, { "epoch": 0.025613314600670054, "grad_norm": 0.34715214371681213, "learning_rate": 8e-05, "loss": 1.6359, "step": 237 }, { "epoch": 0.025721387658056847, "grad_norm": 0.319793164730072, "learning_rate": 8e-05, "loss": 1.524, "step": 238 }, { "epoch": 0.02582946071544364, "grad_norm": 0.37795740365982056, "learning_rate": 8e-05, "loss": 2.0085, "step": 239 }, { "epoch": 0.025937533772830434, "grad_norm": 0.35202550888061523, "learning_rate": 8e-05, "loss": 1.8803, "step": 240 }, { "epoch": 0.026045606830217227, "grad_norm": 0.33381348848342896, "learning_rate": 8e-05, "loss": 1.7462, "step": 241 }, { "epoch": 0.02615367988760402, "grad_norm": 0.36450648307800293, "learning_rate": 8e-05, "loss": 1.9532, "step": 242 }, { "epoch": 0.026261752944990814, "grad_norm": 0.3383338451385498, "learning_rate": 8e-05, "loss": 1.8238, "step": 243 }, { "epoch": 0.026369826002377608, "grad_norm": 0.3532298505306244, "learning_rate": 8e-05, "loss": 1.8886, "step": 244 }, { "epoch": 0.0264778990597644, "grad_norm": 0.3489563763141632, "learning_rate": 8e-05, "loss": 1.8382, "step": 245 }, { "epoch": 0.026585972117151194, "grad_norm": 0.3260352611541748, "learning_rate": 8e-05, "loss": 1.821, "step": 246 }, { "epoch": 0.026694045174537988, "grad_norm": 0.35486161708831787, "learning_rate": 8e-05, "loss": 1.8844, "step": 247 }, { "epoch": 0.02680211823192478, "grad_norm": 0.33174437284469604, "learning_rate": 8e-05, "loss": 1.855, "step": 248 }, { "epoch": 0.026910191289311575, "grad_norm": 0.4141625761985779, "learning_rate": 8e-05, "loss": 1.9501, "step": 249 }, { "epoch": 0.027018264346698368, "grad_norm": 0.37890875339508057, "learning_rate": 8e-05, "loss": 2.0182, "step": 250 }, { "epoch": 0.02712633740408516, "grad_norm": 0.33093369007110596, "learning_rate": 8e-05, "loss": 1.8655, "step": 251 }, { "epoch": 0.027234410461471955, "grad_norm": 0.34627413749694824, "learning_rate": 8e-05, "loss": 1.834, "step": 252 }, { "epoch": 0.02734248351885875, "grad_norm": 0.3526088297367096, "learning_rate": 8e-05, "loss": 1.6507, "step": 253 }, { "epoch": 0.027450556576245542, "grad_norm": 0.3771674931049347, "learning_rate": 8e-05, "loss": 1.9183, "step": 254 }, { "epoch": 0.027558629633632335, "grad_norm": 0.3523033559322357, "learning_rate": 8e-05, "loss": 1.8241, "step": 255 }, { "epoch": 0.02766670269101913, "grad_norm": 0.352531760931015, "learning_rate": 8e-05, "loss": 1.7656, "step": 256 }, { "epoch": 0.027774775748405922, "grad_norm": 0.40698111057281494, "learning_rate": 8e-05, "loss": 2.003, "step": 257 }, { "epoch": 0.027882848805792716, "grad_norm": 0.32587379217147827, "learning_rate": 8e-05, "loss": 1.6962, "step": 258 }, { "epoch": 0.02799092186317951, "grad_norm": 0.37126708030700684, "learning_rate": 8e-05, "loss": 1.9308, "step": 259 }, { "epoch": 0.028098994920566302, "grad_norm": 0.36422014236450195, "learning_rate": 8e-05, "loss": 1.8255, "step": 260 }, { "epoch": 0.028207067977953096, "grad_norm": 0.3457685708999634, "learning_rate": 8e-05, "loss": 1.7153, "step": 261 }, { "epoch": 0.02831514103533989, "grad_norm": 0.35602548718452454, "learning_rate": 8e-05, "loss": 1.8368, "step": 262 }, { "epoch": 0.028423214092726683, "grad_norm": 0.3606073558330536, "learning_rate": 8e-05, "loss": 1.8022, "step": 263 }, { "epoch": 0.028531287150113476, "grad_norm": 0.3583906590938568, "learning_rate": 8e-05, "loss": 1.8543, "step": 264 }, { "epoch": 0.02863936020750027, "grad_norm": 0.3634568750858307, "learning_rate": 8e-05, "loss": 1.7623, "step": 265 }, { "epoch": 0.028747433264887063, "grad_norm": 0.3363524377346039, "learning_rate": 8e-05, "loss": 1.7092, "step": 266 }, { "epoch": 0.028855506322273856, "grad_norm": 0.3427327573299408, "learning_rate": 8e-05, "loss": 1.8065, "step": 267 }, { "epoch": 0.02896357937966065, "grad_norm": 0.36411482095718384, "learning_rate": 8e-05, "loss": 1.6541, "step": 268 }, { "epoch": 0.029071652437047443, "grad_norm": 0.351688027381897, "learning_rate": 8e-05, "loss": 1.7847, "step": 269 }, { "epoch": 0.029179725494434237, "grad_norm": 0.33315518498420715, "learning_rate": 8e-05, "loss": 1.7621, "step": 270 }, { "epoch": 0.02928779855182103, "grad_norm": 0.33381903171539307, "learning_rate": 8e-05, "loss": 1.7731, "step": 271 }, { "epoch": 0.029395871609207824, "grad_norm": 0.3562837541103363, "learning_rate": 8e-05, "loss": 1.8022, "step": 272 }, { "epoch": 0.029503944666594617, "grad_norm": 0.3630593419075012, "learning_rate": 8e-05, "loss": 1.8115, "step": 273 }, { "epoch": 0.02961201772398141, "grad_norm": 0.35505959391593933, "learning_rate": 8e-05, "loss": 1.8505, "step": 274 }, { "epoch": 0.029720090781368204, "grad_norm": 0.3518848121166229, "learning_rate": 8e-05, "loss": 1.799, "step": 275 }, { "epoch": 0.029828163838754997, "grad_norm": 0.3195616900920868, "learning_rate": 8e-05, "loss": 1.7481, "step": 276 }, { "epoch": 0.02993623689614179, "grad_norm": 0.3616880178451538, "learning_rate": 8e-05, "loss": 1.8278, "step": 277 }, { "epoch": 0.030044309953528584, "grad_norm": 0.3618241548538208, "learning_rate": 8e-05, "loss": 1.9174, "step": 278 }, { "epoch": 0.030152383010915378, "grad_norm": 0.3775617182254791, "learning_rate": 8e-05, "loss": 1.979, "step": 279 }, { "epoch": 0.03026045606830217, "grad_norm": 0.33412155508995056, "learning_rate": 8e-05, "loss": 1.8776, "step": 280 }, { "epoch": 0.030368529125688964, "grad_norm": 0.31312301754951477, "learning_rate": 8e-05, "loss": 1.5912, "step": 281 }, { "epoch": 0.030476602183075758, "grad_norm": 0.3662278354167938, "learning_rate": 8e-05, "loss": 1.8465, "step": 282 }, { "epoch": 0.03058467524046255, "grad_norm": 0.36790311336517334, "learning_rate": 8e-05, "loss": 1.9502, "step": 283 }, { "epoch": 0.030692748297849345, "grad_norm": 0.3379440903663635, "learning_rate": 8e-05, "loss": 1.8228, "step": 284 }, { "epoch": 0.030800821355236138, "grad_norm": 0.36950597167015076, "learning_rate": 8e-05, "loss": 1.8648, "step": 285 }, { "epoch": 0.03090889441262293, "grad_norm": 0.3285030424594879, "learning_rate": 8e-05, "loss": 1.6123, "step": 286 }, { "epoch": 0.031016967470009725, "grad_norm": 0.33293578028678894, "learning_rate": 8e-05, "loss": 1.7981, "step": 287 }, { "epoch": 0.03112504052739652, "grad_norm": 0.3433438539505005, "learning_rate": 8e-05, "loss": 1.9394, "step": 288 }, { "epoch": 0.031233113584783312, "grad_norm": 0.3556041717529297, "learning_rate": 8e-05, "loss": 1.9432, "step": 289 }, { "epoch": 0.03134118664217011, "grad_norm": 0.34017595648765564, "learning_rate": 8e-05, "loss": 1.8713, "step": 290 }, { "epoch": 0.0314492596995569, "grad_norm": 0.3288329541683197, "learning_rate": 8e-05, "loss": 1.7685, "step": 291 }, { "epoch": 0.031557332756943696, "grad_norm": 0.35025882720947266, "learning_rate": 8e-05, "loss": 1.6784, "step": 292 }, { "epoch": 0.031665405814330486, "grad_norm": 0.3555113673210144, "learning_rate": 8e-05, "loss": 1.7243, "step": 293 }, { "epoch": 0.03177347887171728, "grad_norm": 0.36543551087379456, "learning_rate": 8e-05, "loss": 1.9049, "step": 294 }, { "epoch": 0.03188155192910407, "grad_norm": 0.3291500210762024, "learning_rate": 8e-05, "loss": 1.6394, "step": 295 }, { "epoch": 0.03198962498649087, "grad_norm": 0.35103124380111694, "learning_rate": 8e-05, "loss": 1.9655, "step": 296 }, { "epoch": 0.03209769804387766, "grad_norm": 0.3578075170516968, "learning_rate": 8e-05, "loss": 1.8393, "step": 297 }, { "epoch": 0.032205771101264456, "grad_norm": 0.3581794798374176, "learning_rate": 8e-05, "loss": 1.7502, "step": 298 }, { "epoch": 0.032313844158651246, "grad_norm": 0.3574570119380951, "learning_rate": 8e-05, "loss": 1.9077, "step": 299 }, { "epoch": 0.03242191721603804, "grad_norm": 0.3462468385696411, "learning_rate": 8e-05, "loss": 1.6018, "step": 300 }, { "epoch": 0.03252999027342483, "grad_norm": 0.37435850501060486, "learning_rate": 8e-05, "loss": 1.9119, "step": 301 }, { "epoch": 0.03263806333081163, "grad_norm": 0.3438180983066559, "learning_rate": 8e-05, "loss": 1.8564, "step": 302 }, { "epoch": 0.03274613638819842, "grad_norm": 0.35696008801460266, "learning_rate": 8e-05, "loss": 1.7753, "step": 303 }, { "epoch": 0.03285420944558522, "grad_norm": 0.3689241409301758, "learning_rate": 8e-05, "loss": 1.6979, "step": 304 }, { "epoch": 0.03296228250297201, "grad_norm": 0.3736790120601654, "learning_rate": 8e-05, "loss": 1.93, "step": 305 }, { "epoch": 0.033070355560358804, "grad_norm": 0.4091819226741791, "learning_rate": 8e-05, "loss": 1.8834, "step": 306 }, { "epoch": 0.033178428617745594, "grad_norm": 0.3624974191188812, "learning_rate": 8e-05, "loss": 1.8233, "step": 307 }, { "epoch": 0.03328650167513239, "grad_norm": 0.40337708592414856, "learning_rate": 8e-05, "loss": 1.91, "step": 308 }, { "epoch": 0.03339457473251918, "grad_norm": 0.3695801794528961, "learning_rate": 8e-05, "loss": 1.8775, "step": 309 }, { "epoch": 0.03350264778990598, "grad_norm": 0.34969937801361084, "learning_rate": 8e-05, "loss": 1.7489, "step": 310 }, { "epoch": 0.03361072084729277, "grad_norm": 0.3449096083641052, "learning_rate": 8e-05, "loss": 1.8207, "step": 311 }, { "epoch": 0.033718793904679564, "grad_norm": 0.32441210746765137, "learning_rate": 8e-05, "loss": 1.5538, "step": 312 }, { "epoch": 0.033826866962066354, "grad_norm": 0.39607810974121094, "learning_rate": 8e-05, "loss": 1.9337, "step": 313 }, { "epoch": 0.03393494001945315, "grad_norm": 0.3832971751689911, "learning_rate": 8e-05, "loss": 2.0217, "step": 314 }, { "epoch": 0.03404301307683994, "grad_norm": 0.36978593468666077, "learning_rate": 8e-05, "loss": 1.9112, "step": 315 }, { "epoch": 0.03415108613422674, "grad_norm": 0.3912685513496399, "learning_rate": 8e-05, "loss": 1.9535, "step": 316 }, { "epoch": 0.03425915919161353, "grad_norm": 0.36194825172424316, "learning_rate": 8e-05, "loss": 1.9126, "step": 317 }, { "epoch": 0.034367232249000325, "grad_norm": 0.3311236798763275, "learning_rate": 8e-05, "loss": 1.786, "step": 318 }, { "epoch": 0.034475305306387115, "grad_norm": 0.35460561513900757, "learning_rate": 8e-05, "loss": 1.8197, "step": 319 }, { "epoch": 0.03458337836377391, "grad_norm": 0.34542927145957947, "learning_rate": 8e-05, "loss": 1.7916, "step": 320 }, { "epoch": 0.0346914514211607, "grad_norm": 0.36881187558174133, "learning_rate": 8e-05, "loss": 1.9422, "step": 321 }, { "epoch": 0.0347995244785475, "grad_norm": 0.32884302735328674, "learning_rate": 8e-05, "loss": 1.8724, "step": 322 }, { "epoch": 0.03490759753593429, "grad_norm": 0.3654920160770416, "learning_rate": 8e-05, "loss": 1.9599, "step": 323 }, { "epoch": 0.035015670593321085, "grad_norm": 0.34346309304237366, "learning_rate": 8e-05, "loss": 1.7999, "step": 324 }, { "epoch": 0.035123743650707875, "grad_norm": 0.346317857503891, "learning_rate": 8e-05, "loss": 1.7251, "step": 325 }, { "epoch": 0.03523181670809467, "grad_norm": 0.3559117317199707, "learning_rate": 8e-05, "loss": 1.9348, "step": 326 }, { "epoch": 0.03533988976548146, "grad_norm": 0.3103870451450348, "learning_rate": 8e-05, "loss": 1.6251, "step": 327 }, { "epoch": 0.03544796282286826, "grad_norm": 0.3160318434238434, "learning_rate": 8e-05, "loss": 1.667, "step": 328 }, { "epoch": 0.03555603588025505, "grad_norm": 0.34836089611053467, "learning_rate": 8e-05, "loss": 1.8578, "step": 329 }, { "epoch": 0.035664108937641846, "grad_norm": 0.34697508811950684, "learning_rate": 8e-05, "loss": 1.8063, "step": 330 }, { "epoch": 0.035772181995028636, "grad_norm": 0.34704625606536865, "learning_rate": 8e-05, "loss": 1.7584, "step": 331 }, { "epoch": 0.03588025505241543, "grad_norm": 0.3697121739387512, "learning_rate": 8e-05, "loss": 1.9914, "step": 332 }, { "epoch": 0.03598832810980222, "grad_norm": 0.3467378616333008, "learning_rate": 8e-05, "loss": 1.7177, "step": 333 }, { "epoch": 0.03609640116718902, "grad_norm": 0.37638208270072937, "learning_rate": 8e-05, "loss": 1.8402, "step": 334 }, { "epoch": 0.03620447422457582, "grad_norm": 0.3592894375324249, "learning_rate": 8e-05, "loss": 1.8893, "step": 335 }, { "epoch": 0.03631254728196261, "grad_norm": 0.3446558713912964, "learning_rate": 8e-05, "loss": 1.8213, "step": 336 }, { "epoch": 0.036420620339349404, "grad_norm": 0.3634584844112396, "learning_rate": 8e-05, "loss": 1.7799, "step": 337 }, { "epoch": 0.036528693396736193, "grad_norm": 0.39033040404319763, "learning_rate": 8e-05, "loss": 1.7989, "step": 338 }, { "epoch": 0.03663676645412299, "grad_norm": 0.4093152582645416, "learning_rate": 8e-05, "loss": 1.9441, "step": 339 }, { "epoch": 0.03674483951150978, "grad_norm": 0.39712831377983093, "learning_rate": 8e-05, "loss": 1.9132, "step": 340 }, { "epoch": 0.03685291256889658, "grad_norm": 0.37408292293548584, "learning_rate": 8e-05, "loss": 1.8234, "step": 341 }, { "epoch": 0.03696098562628337, "grad_norm": 0.35121020674705505, "learning_rate": 8e-05, "loss": 1.6624, "step": 342 }, { "epoch": 0.037069058683670164, "grad_norm": 0.3425823152065277, "learning_rate": 8e-05, "loss": 1.732, "step": 343 }, { "epoch": 0.037177131741056954, "grad_norm": 0.3547641634941101, "learning_rate": 8e-05, "loss": 1.8282, "step": 344 }, { "epoch": 0.03728520479844375, "grad_norm": 0.3267326354980469, "learning_rate": 8e-05, "loss": 1.6574, "step": 345 }, { "epoch": 0.03739327785583054, "grad_norm": 0.37417852878570557, "learning_rate": 8e-05, "loss": 1.9083, "step": 346 }, { "epoch": 0.03750135091321734, "grad_norm": 0.38231778144836426, "learning_rate": 8e-05, "loss": 1.7422, "step": 347 }, { "epoch": 0.03760942397060413, "grad_norm": 0.34262776374816895, "learning_rate": 8e-05, "loss": 1.6912, "step": 348 }, { "epoch": 0.037717497027990925, "grad_norm": 0.3890533149242401, "learning_rate": 8e-05, "loss": 1.8726, "step": 349 }, { "epoch": 0.037825570085377715, "grad_norm": 0.34022581577301025, "learning_rate": 8e-05, "loss": 1.8635, "step": 350 }, { "epoch": 0.03793364314276451, "grad_norm": 0.33824262022972107, "learning_rate": 8e-05, "loss": 1.7204, "step": 351 }, { "epoch": 0.0380417162001513, "grad_norm": 0.3271998167037964, "learning_rate": 8e-05, "loss": 1.5871, "step": 352 }, { "epoch": 0.0381497892575381, "grad_norm": 0.3670596480369568, "learning_rate": 8e-05, "loss": 1.8983, "step": 353 }, { "epoch": 0.03825786231492489, "grad_norm": 0.3383699357509613, "learning_rate": 8e-05, "loss": 1.6058, "step": 354 }, { "epoch": 0.038365935372311685, "grad_norm": 0.34746605157852173, "learning_rate": 8e-05, "loss": 1.692, "step": 355 }, { "epoch": 0.038474008429698475, "grad_norm": 0.3594323694705963, "learning_rate": 8e-05, "loss": 1.7152, "step": 356 }, { "epoch": 0.03858208148708527, "grad_norm": 0.3453253209590912, "learning_rate": 8e-05, "loss": 1.8821, "step": 357 }, { "epoch": 0.03869015454447206, "grad_norm": 0.32800087332725525, "learning_rate": 8e-05, "loss": 1.7831, "step": 358 }, { "epoch": 0.03879822760185886, "grad_norm": 0.35044166445732117, "learning_rate": 8e-05, "loss": 1.7629, "step": 359 }, { "epoch": 0.03890630065924565, "grad_norm": 0.3695552349090576, "learning_rate": 8e-05, "loss": 1.5305, "step": 360 }, { "epoch": 0.039014373716632446, "grad_norm": 0.3436754047870636, "learning_rate": 8e-05, "loss": 1.8017, "step": 361 }, { "epoch": 0.039122446774019236, "grad_norm": 0.41472095251083374, "learning_rate": 8e-05, "loss": 1.3636, "step": 362 }, { "epoch": 0.03923051983140603, "grad_norm": 0.3261834681034088, "learning_rate": 8e-05, "loss": 1.8137, "step": 363 }, { "epoch": 0.03933859288879282, "grad_norm": 0.3855779767036438, "learning_rate": 8e-05, "loss": 1.8365, "step": 364 }, { "epoch": 0.03944666594617962, "grad_norm": 0.361030638217926, "learning_rate": 8e-05, "loss": 1.8141, "step": 365 }, { "epoch": 0.03955473900356641, "grad_norm": 0.3460286855697632, "learning_rate": 8e-05, "loss": 1.6625, "step": 366 }, { "epoch": 0.039662812060953206, "grad_norm": 0.3418448865413666, "learning_rate": 8e-05, "loss": 1.7679, "step": 367 }, { "epoch": 0.039770885118339996, "grad_norm": 0.3445992171764374, "learning_rate": 8e-05, "loss": 1.7121, "step": 368 }, { "epoch": 0.03987895817572679, "grad_norm": 0.3819223940372467, "learning_rate": 8e-05, "loss": 1.7579, "step": 369 }, { "epoch": 0.03998703123311358, "grad_norm": 0.3545958995819092, "learning_rate": 8e-05, "loss": 1.7665, "step": 370 }, { "epoch": 0.04009510429050038, "grad_norm": 0.39416635036468506, "learning_rate": 8e-05, "loss": 1.817, "step": 371 }, { "epoch": 0.04020317734788717, "grad_norm": 0.36554983258247375, "learning_rate": 8e-05, "loss": 1.9476, "step": 372 }, { "epoch": 0.04031125040527397, "grad_norm": 0.3332500159740448, "learning_rate": 8e-05, "loss": 1.8061, "step": 373 }, { "epoch": 0.04041932346266076, "grad_norm": 0.3605744540691376, "learning_rate": 8e-05, "loss": 2.0361, "step": 374 }, { "epoch": 0.040527396520047554, "grad_norm": 0.3399654030799866, "learning_rate": 8e-05, "loss": 1.9372, "step": 375 }, { "epoch": 0.040635469577434344, "grad_norm": 0.3782304525375366, "learning_rate": 8e-05, "loss": 1.8155, "step": 376 }, { "epoch": 0.04074354263482114, "grad_norm": 0.33868125081062317, "learning_rate": 8e-05, "loss": 1.5434, "step": 377 }, { "epoch": 0.04085161569220793, "grad_norm": 0.3886982202529907, "learning_rate": 8e-05, "loss": 1.7462, "step": 378 }, { "epoch": 0.04095968874959473, "grad_norm": 0.4014958441257477, "learning_rate": 8e-05, "loss": 1.8195, "step": 379 }, { "epoch": 0.04106776180698152, "grad_norm": 0.34100309014320374, "learning_rate": 8e-05, "loss": 1.6916, "step": 380 }, { "epoch": 0.041175834864368314, "grad_norm": 0.3505455255508423, "learning_rate": 8e-05, "loss": 1.8154, "step": 381 }, { "epoch": 0.041283907921755104, "grad_norm": 0.3408295810222626, "learning_rate": 8e-05, "loss": 1.79, "step": 382 }, { "epoch": 0.0413919809791419, "grad_norm": 0.3800085484981537, "learning_rate": 8e-05, "loss": 1.8376, "step": 383 }, { "epoch": 0.04150005403652869, "grad_norm": 0.35123538970947266, "learning_rate": 8e-05, "loss": 1.8429, "step": 384 }, { "epoch": 0.04160812709391549, "grad_norm": 0.3258672058582306, "learning_rate": 8e-05, "loss": 1.598, "step": 385 }, { "epoch": 0.04171620015130228, "grad_norm": 0.3586743175983429, "learning_rate": 8e-05, "loss": 1.8037, "step": 386 }, { "epoch": 0.041824273208689075, "grad_norm": 0.3811790645122528, "learning_rate": 8e-05, "loss": 1.7988, "step": 387 }, { "epoch": 0.041932346266075865, "grad_norm": 0.36081990599632263, "learning_rate": 8e-05, "loss": 1.8521, "step": 388 }, { "epoch": 0.04204041932346266, "grad_norm": 0.35482174158096313, "learning_rate": 8e-05, "loss": 1.6141, "step": 389 }, { "epoch": 0.04214849238084945, "grad_norm": 0.3619999885559082, "learning_rate": 8e-05, "loss": 1.6408, "step": 390 }, { "epoch": 0.04225656543823625, "grad_norm": 0.371709942817688, "learning_rate": 8e-05, "loss": 1.9687, "step": 391 }, { "epoch": 0.04236463849562304, "grad_norm": 0.32231834530830383, "learning_rate": 8e-05, "loss": 1.5901, "step": 392 }, { "epoch": 0.042472711553009836, "grad_norm": 0.33641085028648376, "learning_rate": 8e-05, "loss": 1.7405, "step": 393 }, { "epoch": 0.042580784610396626, "grad_norm": 0.35316988825798035, "learning_rate": 8e-05, "loss": 1.7697, "step": 394 }, { "epoch": 0.04268885766778342, "grad_norm": 0.3627004623413086, "learning_rate": 8e-05, "loss": 1.8328, "step": 395 }, { "epoch": 0.04279693072517021, "grad_norm": 0.3517940044403076, "learning_rate": 8e-05, "loss": 1.8157, "step": 396 }, { "epoch": 0.04290500378255701, "grad_norm": 0.34521055221557617, "learning_rate": 8e-05, "loss": 1.9056, "step": 397 }, { "epoch": 0.0430130768399438, "grad_norm": 0.3528774380683899, "learning_rate": 8e-05, "loss": 1.7603, "step": 398 }, { "epoch": 0.043121149897330596, "grad_norm": 0.32668769359588623, "learning_rate": 8e-05, "loss": 1.7195, "step": 399 }, { "epoch": 0.043229222954717386, "grad_norm": 0.36554402112960815, "learning_rate": 8e-05, "loss": 1.8537, "step": 400 }, { "epoch": 0.04333729601210418, "grad_norm": 0.34430208802223206, "learning_rate": 8e-05, "loss": 1.8433, "step": 401 }, { "epoch": 0.04344536906949097, "grad_norm": 0.35406693816185, "learning_rate": 8e-05, "loss": 1.8906, "step": 402 }, { "epoch": 0.04355344212687777, "grad_norm": 0.3372669816017151, "learning_rate": 8e-05, "loss": 1.7439, "step": 403 }, { "epoch": 0.04366151518426456, "grad_norm": 0.3269980251789093, "learning_rate": 8e-05, "loss": 1.7987, "step": 404 }, { "epoch": 0.04376958824165136, "grad_norm": 0.35388675332069397, "learning_rate": 8e-05, "loss": 1.8383, "step": 405 }, { "epoch": 0.04387766129903815, "grad_norm": 0.3406014144420624, "learning_rate": 8e-05, "loss": 1.8139, "step": 406 }, { "epoch": 0.043985734356424944, "grad_norm": 0.3585246801376343, "learning_rate": 8e-05, "loss": 1.7213, "step": 407 }, { "epoch": 0.044093807413811734, "grad_norm": 0.38654810190200806, "learning_rate": 8e-05, "loss": 1.8619, "step": 408 }, { "epoch": 0.04420188047119853, "grad_norm": 0.38541606068611145, "learning_rate": 8e-05, "loss": 1.8839, "step": 409 }, { "epoch": 0.04430995352858532, "grad_norm": 0.34015166759490967, "learning_rate": 8e-05, "loss": 1.8117, "step": 410 }, { "epoch": 0.04441802658597212, "grad_norm": 0.37512245774269104, "learning_rate": 8e-05, "loss": 1.8673, "step": 411 }, { "epoch": 0.04452609964335891, "grad_norm": 0.3188580870628357, "learning_rate": 8e-05, "loss": 1.612, "step": 412 }, { "epoch": 0.044634172700745704, "grad_norm": 0.3648746907711029, "learning_rate": 8e-05, "loss": 1.8527, "step": 413 }, { "epoch": 0.044742245758132494, "grad_norm": 0.45930737257003784, "learning_rate": 8e-05, "loss": 1.8655, "step": 414 }, { "epoch": 0.04485031881551929, "grad_norm": 0.3602280914783478, "learning_rate": 8e-05, "loss": 1.8813, "step": 415 }, { "epoch": 0.04495839187290608, "grad_norm": 0.34455743432044983, "learning_rate": 8e-05, "loss": 1.8455, "step": 416 }, { "epoch": 0.04506646493029288, "grad_norm": 0.331175833940506, "learning_rate": 8e-05, "loss": 1.7908, "step": 417 }, { "epoch": 0.045174537987679675, "grad_norm": 0.33429694175720215, "learning_rate": 8e-05, "loss": 1.6788, "step": 418 }, { "epoch": 0.045282611045066465, "grad_norm": 0.374340683221817, "learning_rate": 8e-05, "loss": 1.923, "step": 419 }, { "epoch": 0.04539068410245326, "grad_norm": 0.37301379442214966, "learning_rate": 8e-05, "loss": 1.6682, "step": 420 }, { "epoch": 0.04549875715984005, "grad_norm": 0.3516581952571869, "learning_rate": 8e-05, "loss": 1.7969, "step": 421 }, { "epoch": 0.04560683021722685, "grad_norm": 0.34179389476776123, "learning_rate": 8e-05, "loss": 1.799, "step": 422 }, { "epoch": 0.04571490327461364, "grad_norm": 0.34728938341140747, "learning_rate": 8e-05, "loss": 1.8193, "step": 423 }, { "epoch": 0.045822976332000435, "grad_norm": 0.3493211567401886, "learning_rate": 8e-05, "loss": 1.7621, "step": 424 }, { "epoch": 0.045931049389387225, "grad_norm": 0.34878382086753845, "learning_rate": 8e-05, "loss": 1.8856, "step": 425 }, { "epoch": 0.04603912244677402, "grad_norm": 0.3423656225204468, "learning_rate": 8e-05, "loss": 1.8972, "step": 426 }, { "epoch": 0.04614719550416081, "grad_norm": 0.36060187220573425, "learning_rate": 8e-05, "loss": 1.7659, "step": 427 }, { "epoch": 0.04625526856154761, "grad_norm": 0.36862578988075256, "learning_rate": 8e-05, "loss": 1.848, "step": 428 }, { "epoch": 0.0463633416189344, "grad_norm": 0.3442145586013794, "learning_rate": 8e-05, "loss": 1.7151, "step": 429 }, { "epoch": 0.046471414676321196, "grad_norm": 0.3400770425796509, "learning_rate": 8e-05, "loss": 1.852, "step": 430 }, { "epoch": 0.046579487733707986, "grad_norm": 0.3153742849826813, "learning_rate": 8e-05, "loss": 1.8187, "step": 431 }, { "epoch": 0.04668756079109478, "grad_norm": 0.36666610836982727, "learning_rate": 8e-05, "loss": 1.732, "step": 432 }, { "epoch": 0.04679563384848157, "grad_norm": 0.35173097252845764, "learning_rate": 8e-05, "loss": 1.794, "step": 433 }, { "epoch": 0.04690370690586837, "grad_norm": 0.3420298397541046, "learning_rate": 8e-05, "loss": 1.8638, "step": 434 }, { "epoch": 0.04701177996325516, "grad_norm": 0.34960559010505676, "learning_rate": 8e-05, "loss": 1.8717, "step": 435 }, { "epoch": 0.04711985302064196, "grad_norm": 0.3430729806423187, "learning_rate": 8e-05, "loss": 1.7593, "step": 436 }, { "epoch": 0.04722792607802875, "grad_norm": 0.3420880436897278, "learning_rate": 8e-05, "loss": 1.7324, "step": 437 }, { "epoch": 0.047335999135415543, "grad_norm": 0.38300836086273193, "learning_rate": 8e-05, "loss": 1.9481, "step": 438 }, { "epoch": 0.04744407219280233, "grad_norm": 0.3362177014350891, "learning_rate": 8e-05, "loss": 1.8758, "step": 439 }, { "epoch": 0.04755214525018913, "grad_norm": 0.3980714678764343, "learning_rate": 8e-05, "loss": 1.9653, "step": 440 }, { "epoch": 0.04766021830757592, "grad_norm": 0.38389119505882263, "learning_rate": 8e-05, "loss": 1.912, "step": 441 }, { "epoch": 0.04776829136496272, "grad_norm": 0.36654260754585266, "learning_rate": 8e-05, "loss": 1.8661, "step": 442 }, { "epoch": 0.04787636442234951, "grad_norm": 0.3643897473812103, "learning_rate": 8e-05, "loss": 1.7236, "step": 443 }, { "epoch": 0.047984437479736304, "grad_norm": 0.33685579895973206, "learning_rate": 8e-05, "loss": 1.7151, "step": 444 }, { "epoch": 0.048092510537123094, "grad_norm": 0.3855327069759369, "learning_rate": 8e-05, "loss": 1.8651, "step": 445 }, { "epoch": 0.04820058359450989, "grad_norm": 0.37553495168685913, "learning_rate": 8e-05, "loss": 1.8213, "step": 446 }, { "epoch": 0.04830865665189668, "grad_norm": 0.3785769045352936, "learning_rate": 8e-05, "loss": 1.5616, "step": 447 }, { "epoch": 0.04841672970928348, "grad_norm": 0.36590754985809326, "learning_rate": 8e-05, "loss": 1.7161, "step": 448 }, { "epoch": 0.04852480276667027, "grad_norm": 0.3443998396396637, "learning_rate": 8e-05, "loss": 1.5348, "step": 449 }, { "epoch": 0.048632875824057065, "grad_norm": 0.37892866134643555, "learning_rate": 8e-05, "loss": 1.959, "step": 450 }, { "epoch": 0.048740948881443855, "grad_norm": 0.35763201117515564, "learning_rate": 8e-05, "loss": 1.836, "step": 451 }, { "epoch": 0.04884902193883065, "grad_norm": 0.33352231979370117, "learning_rate": 8e-05, "loss": 1.7613, "step": 452 }, { "epoch": 0.04895709499621744, "grad_norm": 0.3498244285583496, "learning_rate": 8e-05, "loss": 1.7944, "step": 453 }, { "epoch": 0.04906516805360424, "grad_norm": 0.37402021884918213, "learning_rate": 8e-05, "loss": 1.935, "step": 454 }, { "epoch": 0.04917324111099103, "grad_norm": 0.3497718274593353, "learning_rate": 8e-05, "loss": 1.7446, "step": 455 }, { "epoch": 0.049281314168377825, "grad_norm": 0.4347609281539917, "learning_rate": 8e-05, "loss": 1.5846, "step": 456 }, { "epoch": 0.049389387225764615, "grad_norm": 0.374033123254776, "learning_rate": 8e-05, "loss": 2.0395, "step": 457 }, { "epoch": 0.04949746028315141, "grad_norm": 0.3523750305175781, "learning_rate": 8e-05, "loss": 1.9116, "step": 458 }, { "epoch": 0.0496055333405382, "grad_norm": 0.3229548931121826, "learning_rate": 8e-05, "loss": 1.6511, "step": 459 }, { "epoch": 0.049713606397925, "grad_norm": 0.3338475227355957, "learning_rate": 8e-05, "loss": 1.8061, "step": 460 }, { "epoch": 0.04982167945531179, "grad_norm": 0.35802769660949707, "learning_rate": 8e-05, "loss": 1.76, "step": 461 }, { "epoch": 0.049929752512698586, "grad_norm": 0.3635770380496979, "learning_rate": 8e-05, "loss": 1.8359, "step": 462 }, { "epoch": 0.050037825570085376, "grad_norm": 0.3167588412761688, "learning_rate": 8e-05, "loss": 1.5331, "step": 463 }, { "epoch": 0.05014589862747217, "grad_norm": 0.3561248481273651, "learning_rate": 8e-05, "loss": 1.6975, "step": 464 }, { "epoch": 0.05025397168485896, "grad_norm": 0.33736878633499146, "learning_rate": 8e-05, "loss": 1.7406, "step": 465 }, { "epoch": 0.05036204474224576, "grad_norm": 0.3916040062904358, "learning_rate": 8e-05, "loss": 2.002, "step": 466 }, { "epoch": 0.05047011779963255, "grad_norm": 0.33673620223999023, "learning_rate": 8e-05, "loss": 1.6607, "step": 467 }, { "epoch": 0.050578190857019346, "grad_norm": 0.35699862241744995, "learning_rate": 8e-05, "loss": 1.8892, "step": 468 }, { "epoch": 0.050686263914406136, "grad_norm": 0.34727856516838074, "learning_rate": 8e-05, "loss": 1.7731, "step": 469 }, { "epoch": 0.05079433697179293, "grad_norm": 0.32350873947143555, "learning_rate": 8e-05, "loss": 1.7876, "step": 470 }, { "epoch": 0.05090241002917972, "grad_norm": 0.37469542026519775, "learning_rate": 8e-05, "loss": 1.965, "step": 471 }, { "epoch": 0.05101048308656652, "grad_norm": 0.37748417258262634, "learning_rate": 8e-05, "loss": 1.7888, "step": 472 }, { "epoch": 0.05111855614395331, "grad_norm": 0.35077032446861267, "learning_rate": 8e-05, "loss": 1.8116, "step": 473 }, { "epoch": 0.05122662920134011, "grad_norm": 0.36138689517974854, "learning_rate": 8e-05, "loss": 1.7981, "step": 474 }, { "epoch": 0.0513347022587269, "grad_norm": 0.37589678168296814, "learning_rate": 8e-05, "loss": 2.1543, "step": 475 }, { "epoch": 0.051442775316113694, "grad_norm": 0.37126532196998596, "learning_rate": 8e-05, "loss": 1.8198, "step": 476 }, { "epoch": 0.051550848373500484, "grad_norm": 0.3406098484992981, "learning_rate": 8e-05, "loss": 1.7297, "step": 477 }, { "epoch": 0.05165892143088728, "grad_norm": 0.353903591632843, "learning_rate": 8e-05, "loss": 1.9113, "step": 478 }, { "epoch": 0.05176699448827407, "grad_norm": 0.342562198638916, "learning_rate": 8e-05, "loss": 1.7864, "step": 479 }, { "epoch": 0.05187506754566087, "grad_norm": 0.34356650710105896, "learning_rate": 8e-05, "loss": 1.7997, "step": 480 }, { "epoch": 0.05198314060304766, "grad_norm": 0.32555460929870605, "learning_rate": 8e-05, "loss": 1.6289, "step": 481 }, { "epoch": 0.052091213660434454, "grad_norm": 0.37806424498558044, "learning_rate": 8e-05, "loss": 1.7397, "step": 482 }, { "epoch": 0.052199286717821244, "grad_norm": 0.40518635511398315, "learning_rate": 8e-05, "loss": 1.8266, "step": 483 }, { "epoch": 0.05230735977520804, "grad_norm": 0.35818538069725037, "learning_rate": 8e-05, "loss": 1.6793, "step": 484 }, { "epoch": 0.05241543283259483, "grad_norm": 0.3218657076358795, "learning_rate": 8e-05, "loss": 1.7153, "step": 485 }, { "epoch": 0.05252350588998163, "grad_norm": 0.40635812282562256, "learning_rate": 8e-05, "loss": 1.761, "step": 486 }, { "epoch": 0.05263157894736842, "grad_norm": 0.39171960949897766, "learning_rate": 8e-05, "loss": 1.5384, "step": 487 }, { "epoch": 0.052739652004755215, "grad_norm": 0.3521215319633484, "learning_rate": 8e-05, "loss": 1.7272, "step": 488 }, { "epoch": 0.052847725062142005, "grad_norm": 0.3266386389732361, "learning_rate": 8e-05, "loss": 1.6441, "step": 489 }, { "epoch": 0.0529557981195288, "grad_norm": 0.3504844605922699, "learning_rate": 8e-05, "loss": 1.7481, "step": 490 }, { "epoch": 0.05306387117691559, "grad_norm": 0.389176607131958, "learning_rate": 8e-05, "loss": 1.9254, "step": 491 }, { "epoch": 0.05317194423430239, "grad_norm": 0.36079034209251404, "learning_rate": 8e-05, "loss": 1.7437, "step": 492 }, { "epoch": 0.05328001729168918, "grad_norm": 0.37580999732017517, "learning_rate": 8e-05, "loss": 1.9303, "step": 493 }, { "epoch": 0.053388090349075976, "grad_norm": 0.37816786766052246, "learning_rate": 8e-05, "loss": 1.8253, "step": 494 }, { "epoch": 0.053496163406462766, "grad_norm": 0.3468021750450134, "learning_rate": 8e-05, "loss": 1.7591, "step": 495 }, { "epoch": 0.05360423646384956, "grad_norm": 0.37619641423225403, "learning_rate": 8e-05, "loss": 1.8257, "step": 496 }, { "epoch": 0.05371230952123635, "grad_norm": 0.34040290117263794, "learning_rate": 8e-05, "loss": 1.9514, "step": 497 }, { "epoch": 0.05382038257862315, "grad_norm": 0.3603047728538513, "learning_rate": 8e-05, "loss": 1.8091, "step": 498 }, { "epoch": 0.05392845563600994, "grad_norm": 0.3678261637687683, "learning_rate": 8e-05, "loss": 1.8713, "step": 499 }, { "epoch": 0.054036528693396736, "grad_norm": 0.32692885398864746, "learning_rate": 8e-05, "loss": 1.7825, "step": 500 }, { "epoch": 0.05414460175078353, "grad_norm": 0.33240586519241333, "learning_rate": 8e-05, "loss": 1.7369, "step": 501 }, { "epoch": 0.05425267480817032, "grad_norm": 0.3406820297241211, "learning_rate": 8e-05, "loss": 1.9478, "step": 502 }, { "epoch": 0.05436074786555712, "grad_norm": 0.3288944959640503, "learning_rate": 8e-05, "loss": 1.6793, "step": 503 }, { "epoch": 0.05446882092294391, "grad_norm": 0.3384639620780945, "learning_rate": 8e-05, "loss": 1.9113, "step": 504 }, { "epoch": 0.05457689398033071, "grad_norm": 0.341382771730423, "learning_rate": 8e-05, "loss": 1.9594, "step": 505 }, { "epoch": 0.0546849670377175, "grad_norm": 0.338291198015213, "learning_rate": 8e-05, "loss": 1.78, "step": 506 }, { "epoch": 0.054793040095104294, "grad_norm": 0.3306415379047394, "learning_rate": 8e-05, "loss": 1.7005, "step": 507 }, { "epoch": 0.054901113152491084, "grad_norm": 0.33666789531707764, "learning_rate": 8e-05, "loss": 1.6774, "step": 508 }, { "epoch": 0.05500918620987788, "grad_norm": 0.3264491856098175, "learning_rate": 8e-05, "loss": 1.7859, "step": 509 }, { "epoch": 0.05511725926726467, "grad_norm": 0.3928073048591614, "learning_rate": 8e-05, "loss": 1.8382, "step": 510 }, { "epoch": 0.05522533232465147, "grad_norm": 0.3506225347518921, "learning_rate": 8e-05, "loss": 1.9297, "step": 511 }, { "epoch": 0.05533340538203826, "grad_norm": 0.3470093607902527, "learning_rate": 8e-05, "loss": 1.7875, "step": 512 }, { "epoch": 0.055441478439425054, "grad_norm": 0.3335685729980469, "learning_rate": 8e-05, "loss": 1.8113, "step": 513 }, { "epoch": 0.055549551496811844, "grad_norm": 0.3314863443374634, "learning_rate": 8e-05, "loss": 1.7284, "step": 514 }, { "epoch": 0.05565762455419864, "grad_norm": 0.33801519870758057, "learning_rate": 8e-05, "loss": 1.7993, "step": 515 }, { "epoch": 0.05576569761158543, "grad_norm": 0.3557450771331787, "learning_rate": 8e-05, "loss": 1.6081, "step": 516 }, { "epoch": 0.05587377066897223, "grad_norm": 0.4297201335430145, "learning_rate": 8e-05, "loss": 2.0665, "step": 517 }, { "epoch": 0.05598184372635902, "grad_norm": 0.35107314586639404, "learning_rate": 8e-05, "loss": 1.9051, "step": 518 }, { "epoch": 0.056089916783745815, "grad_norm": 0.3304230868816376, "learning_rate": 8e-05, "loss": 1.7827, "step": 519 }, { "epoch": 0.056197989841132605, "grad_norm": 0.35226425528526306, "learning_rate": 8e-05, "loss": 1.7096, "step": 520 }, { "epoch": 0.0563060628985194, "grad_norm": 0.32998597621917725, "learning_rate": 8e-05, "loss": 1.7804, "step": 521 }, { "epoch": 0.05641413595590619, "grad_norm": 0.3704870939254761, "learning_rate": 8e-05, "loss": 1.8446, "step": 522 }, { "epoch": 0.05652220901329299, "grad_norm": 0.3906189501285553, "learning_rate": 8e-05, "loss": 1.8716, "step": 523 }, { "epoch": 0.05663028207067978, "grad_norm": 0.3609927296638489, "learning_rate": 8e-05, "loss": 1.8022, "step": 524 }, { "epoch": 0.056738355128066575, "grad_norm": 0.3408806025981903, "learning_rate": 8e-05, "loss": 1.7892, "step": 525 }, { "epoch": 0.056846428185453365, "grad_norm": 0.3482629656791687, "learning_rate": 8e-05, "loss": 1.7619, "step": 526 }, { "epoch": 0.05695450124284016, "grad_norm": 0.34368976950645447, "learning_rate": 8e-05, "loss": 1.7962, "step": 527 }, { "epoch": 0.05706257430022695, "grad_norm": 0.35801342129707336, "learning_rate": 8e-05, "loss": 1.7472, "step": 528 }, { "epoch": 0.05717064735761375, "grad_norm": 0.3331338167190552, "learning_rate": 8e-05, "loss": 1.7631, "step": 529 }, { "epoch": 0.05727872041500054, "grad_norm": 0.3434489965438843, "learning_rate": 8e-05, "loss": 1.7431, "step": 530 }, { "epoch": 0.057386793472387336, "grad_norm": 0.3381546139717102, "learning_rate": 8e-05, "loss": 1.7641, "step": 531 }, { "epoch": 0.057494866529774126, "grad_norm": 0.33945730328559875, "learning_rate": 8e-05, "loss": 1.9493, "step": 532 }, { "epoch": 0.05760293958716092, "grad_norm": 0.3365679383277893, "learning_rate": 8e-05, "loss": 1.8159, "step": 533 }, { "epoch": 0.05771101264454771, "grad_norm": 0.3401789963245392, "learning_rate": 8e-05, "loss": 1.8287, "step": 534 }, { "epoch": 0.05781908570193451, "grad_norm": 0.33047136664390564, "learning_rate": 8e-05, "loss": 1.7604, "step": 535 }, { "epoch": 0.0579271587593213, "grad_norm": 0.35353079438209534, "learning_rate": 8e-05, "loss": 1.9169, "step": 536 }, { "epoch": 0.0580352318167081, "grad_norm": 0.39713647961616516, "learning_rate": 8e-05, "loss": 1.9314, "step": 537 }, { "epoch": 0.05814330487409489, "grad_norm": 0.3830745816230774, "learning_rate": 8e-05, "loss": 1.9128, "step": 538 }, { "epoch": 0.058251377931481683, "grad_norm": 0.36057743430137634, "learning_rate": 8e-05, "loss": 1.7856, "step": 539 }, { "epoch": 0.05835945098886847, "grad_norm": 0.40052130818367004, "learning_rate": 8e-05, "loss": 1.9473, "step": 540 }, { "epoch": 0.05846752404625527, "grad_norm": 0.340400367975235, "learning_rate": 8e-05, "loss": 1.7526, "step": 541 }, { "epoch": 0.05857559710364206, "grad_norm": 0.334179550409317, "learning_rate": 8e-05, "loss": 1.6496, "step": 542 }, { "epoch": 0.05868367016102886, "grad_norm": 0.3954176902770996, "learning_rate": 8e-05, "loss": 1.7186, "step": 543 }, { "epoch": 0.05879174321841565, "grad_norm": 0.40544360876083374, "learning_rate": 8e-05, "loss": 1.9036, "step": 544 }, { "epoch": 0.058899816275802444, "grad_norm": 0.3730384409427643, "learning_rate": 8e-05, "loss": 1.9688, "step": 545 }, { "epoch": 0.059007889333189234, "grad_norm": 0.38895460963249207, "learning_rate": 8e-05, "loss": 1.8152, "step": 546 }, { "epoch": 0.05911596239057603, "grad_norm": 0.3750901520252228, "learning_rate": 8e-05, "loss": 1.92, "step": 547 }, { "epoch": 0.05922403544796282, "grad_norm": 0.3438848555088043, "learning_rate": 8e-05, "loss": 1.8058, "step": 548 }, { "epoch": 0.05933210850534962, "grad_norm": 0.34970834851264954, "learning_rate": 8e-05, "loss": 1.7938, "step": 549 }, { "epoch": 0.05944018156273641, "grad_norm": 0.37738972902297974, "learning_rate": 8e-05, "loss": 2.0443, "step": 550 }, { "epoch": 0.059548254620123205, "grad_norm": 0.3614276647567749, "learning_rate": 8e-05, "loss": 1.8913, "step": 551 }, { "epoch": 0.059656327677509995, "grad_norm": 0.36763495206832886, "learning_rate": 8e-05, "loss": 1.8611, "step": 552 }, { "epoch": 0.05976440073489679, "grad_norm": 0.35178446769714355, "learning_rate": 8e-05, "loss": 1.8826, "step": 553 }, { "epoch": 0.05987247379228358, "grad_norm": 0.3617529273033142, "learning_rate": 8e-05, "loss": 1.8131, "step": 554 }, { "epoch": 0.05998054684967038, "grad_norm": 0.35756927728652954, "learning_rate": 8e-05, "loss": 1.7225, "step": 555 }, { "epoch": 0.06008861990705717, "grad_norm": 0.3592441976070404, "learning_rate": 8e-05, "loss": 1.7632, "step": 556 }, { "epoch": 0.060196692964443965, "grad_norm": 0.39115583896636963, "learning_rate": 8e-05, "loss": 2.0189, "step": 557 }, { "epoch": 0.060304766021830755, "grad_norm": 0.3478936553001404, "learning_rate": 8e-05, "loss": 1.911, "step": 558 }, { "epoch": 0.06041283907921755, "grad_norm": 0.382700115442276, "learning_rate": 8e-05, "loss": 1.7844, "step": 559 }, { "epoch": 0.06052091213660434, "grad_norm": 0.33886781334877014, "learning_rate": 8e-05, "loss": 1.7133, "step": 560 }, { "epoch": 0.06062898519399114, "grad_norm": 0.32599201798439026, "learning_rate": 8e-05, "loss": 1.7179, "step": 561 }, { "epoch": 0.06073705825137793, "grad_norm": 0.3453141450881958, "learning_rate": 8e-05, "loss": 1.8782, "step": 562 }, { "epoch": 0.060845131308764726, "grad_norm": 0.3136008083820343, "learning_rate": 8e-05, "loss": 1.8128, "step": 563 }, { "epoch": 0.060953204366151516, "grad_norm": 0.34391477704048157, "learning_rate": 8e-05, "loss": 1.5151, "step": 564 }, { "epoch": 0.06106127742353831, "grad_norm": 0.3599853217601776, "learning_rate": 8e-05, "loss": 1.9333, "step": 565 }, { "epoch": 0.0611693504809251, "grad_norm": 0.3889324963092804, "learning_rate": 8e-05, "loss": 2.0541, "step": 566 }, { "epoch": 0.0612774235383119, "grad_norm": 0.33915358781814575, "learning_rate": 8e-05, "loss": 1.8674, "step": 567 }, { "epoch": 0.06138549659569869, "grad_norm": 0.36452826857566833, "learning_rate": 8e-05, "loss": 1.8277, "step": 568 }, { "epoch": 0.061493569653085486, "grad_norm": 0.34183764457702637, "learning_rate": 8e-05, "loss": 1.865, "step": 569 }, { "epoch": 0.061601642710472276, "grad_norm": 0.39172831177711487, "learning_rate": 8e-05, "loss": 1.9538, "step": 570 }, { "epoch": 0.06170971576785907, "grad_norm": 0.3603333532810211, "learning_rate": 8e-05, "loss": 1.9699, "step": 571 }, { "epoch": 0.06181778882524586, "grad_norm": 0.3480854332447052, "learning_rate": 8e-05, "loss": 1.8306, "step": 572 }, { "epoch": 0.06192586188263266, "grad_norm": 0.34380054473876953, "learning_rate": 8e-05, "loss": 1.783, "step": 573 }, { "epoch": 0.06203393494001945, "grad_norm": 0.33279845118522644, "learning_rate": 8e-05, "loss": 1.7465, "step": 574 }, { "epoch": 0.06214200799740625, "grad_norm": 0.3506140410900116, "learning_rate": 8e-05, "loss": 1.8859, "step": 575 }, { "epoch": 0.06225008105479304, "grad_norm": 0.34136146306991577, "learning_rate": 8e-05, "loss": 1.7201, "step": 576 }, { "epoch": 0.062358154112179834, "grad_norm": 0.37575459480285645, "learning_rate": 8e-05, "loss": 1.8205, "step": 577 }, { "epoch": 0.062466227169566624, "grad_norm": 0.3390471637248993, "learning_rate": 8e-05, "loss": 1.6908, "step": 578 }, { "epoch": 0.06257430022695341, "grad_norm": 0.37138631939888, "learning_rate": 8e-05, "loss": 1.9054, "step": 579 }, { "epoch": 0.06268237328434022, "grad_norm": 0.32744237780570984, "learning_rate": 8e-05, "loss": 1.7409, "step": 580 }, { "epoch": 0.06279044634172701, "grad_norm": 0.32661354541778564, "learning_rate": 8e-05, "loss": 1.5193, "step": 581 }, { "epoch": 0.0628985193991138, "grad_norm": 0.3683415353298187, "learning_rate": 8e-05, "loss": 1.9487, "step": 582 }, { "epoch": 0.06300659245650059, "grad_norm": 0.36816683411598206, "learning_rate": 8e-05, "loss": 1.7186, "step": 583 }, { "epoch": 0.06311466551388739, "grad_norm": 0.34831303358078003, "learning_rate": 8e-05, "loss": 1.857, "step": 584 }, { "epoch": 0.06322273857127418, "grad_norm": 0.3598869740962982, "learning_rate": 8e-05, "loss": 1.7863, "step": 585 }, { "epoch": 0.06333081162866097, "grad_norm": 0.34880056977272034, "learning_rate": 8e-05, "loss": 1.7589, "step": 586 }, { "epoch": 0.06343888468604778, "grad_norm": 0.3511759638786316, "learning_rate": 8e-05, "loss": 1.6893, "step": 587 }, { "epoch": 0.06354695774343457, "grad_norm": 0.35798949003219604, "learning_rate": 8e-05, "loss": 1.7675, "step": 588 }, { "epoch": 0.06365503080082136, "grad_norm": 0.38478884100914, "learning_rate": 8e-05, "loss": 1.9547, "step": 589 }, { "epoch": 0.06376310385820814, "grad_norm": 0.3490166664123535, "learning_rate": 8e-05, "loss": 1.9554, "step": 590 }, { "epoch": 0.06387117691559495, "grad_norm": 0.33670371770858765, "learning_rate": 8e-05, "loss": 1.8227, "step": 591 }, { "epoch": 0.06397924997298174, "grad_norm": 0.3731778562068939, "learning_rate": 8e-05, "loss": 1.7479, "step": 592 }, { "epoch": 0.06408732303036853, "grad_norm": 0.36261558532714844, "learning_rate": 8e-05, "loss": 1.9185, "step": 593 }, { "epoch": 0.06419539608775532, "grad_norm": 0.341047465801239, "learning_rate": 8e-05, "loss": 1.7979, "step": 594 }, { "epoch": 0.06430346914514212, "grad_norm": 0.32539796829223633, "learning_rate": 8e-05, "loss": 1.6979, "step": 595 }, { "epoch": 0.06441154220252891, "grad_norm": 0.347179114818573, "learning_rate": 8e-05, "loss": 1.8035, "step": 596 }, { "epoch": 0.0645196152599157, "grad_norm": 0.3422594368457794, "learning_rate": 8e-05, "loss": 1.6975, "step": 597 }, { "epoch": 0.06462768831730249, "grad_norm": 0.3355692923069, "learning_rate": 8e-05, "loss": 1.7896, "step": 598 }, { "epoch": 0.0647357613746893, "grad_norm": 0.36092662811279297, "learning_rate": 8e-05, "loss": 1.8366, "step": 599 }, { "epoch": 0.06484383443207609, "grad_norm": 0.37264716625213623, "learning_rate": 8e-05, "loss": 1.8769, "step": 600 }, { "epoch": 0.06495190748946288, "grad_norm": 0.33752110600471497, "learning_rate": 8e-05, "loss": 1.5776, "step": 601 }, { "epoch": 0.06505998054684967, "grad_norm": 0.3596280813217163, "learning_rate": 8e-05, "loss": 1.8005, "step": 602 }, { "epoch": 0.06516805360423647, "grad_norm": 0.34022724628448486, "learning_rate": 8e-05, "loss": 1.7063, "step": 603 }, { "epoch": 0.06527612666162326, "grad_norm": 0.3194156587123871, "learning_rate": 8e-05, "loss": 1.6595, "step": 604 }, { "epoch": 0.06538419971901005, "grad_norm": 0.3691875636577606, "learning_rate": 8e-05, "loss": 1.8642, "step": 605 }, { "epoch": 0.06549227277639684, "grad_norm": 0.34021759033203125, "learning_rate": 8e-05, "loss": 1.6751, "step": 606 }, { "epoch": 0.06560034583378364, "grad_norm": 0.42838627099990845, "learning_rate": 8e-05, "loss": 2.0069, "step": 607 }, { "epoch": 0.06570841889117043, "grad_norm": 0.355054646730423, "learning_rate": 8e-05, "loss": 1.7867, "step": 608 }, { "epoch": 0.06581649194855722, "grad_norm": 0.3429180681705475, "learning_rate": 8e-05, "loss": 1.5826, "step": 609 }, { "epoch": 0.06592456500594401, "grad_norm": 0.3588315546512604, "learning_rate": 8e-05, "loss": 1.7159, "step": 610 }, { "epoch": 0.06603263806333082, "grad_norm": 0.33538955450057983, "learning_rate": 8e-05, "loss": 1.6732, "step": 611 }, { "epoch": 0.06614071112071761, "grad_norm": 0.345381498336792, "learning_rate": 8e-05, "loss": 1.7575, "step": 612 }, { "epoch": 0.0662487841781044, "grad_norm": 0.3466568887233734, "learning_rate": 8e-05, "loss": 1.7002, "step": 613 }, { "epoch": 0.06635685723549119, "grad_norm": 0.3260331153869629, "learning_rate": 8e-05, "loss": 1.7258, "step": 614 }, { "epoch": 0.06646493029287799, "grad_norm": 0.3282583951950073, "learning_rate": 8e-05, "loss": 1.8235, "step": 615 }, { "epoch": 0.06657300335026478, "grad_norm": 0.37326928973197937, "learning_rate": 8e-05, "loss": 1.9629, "step": 616 }, { "epoch": 0.06668107640765157, "grad_norm": 0.36559388041496277, "learning_rate": 8e-05, "loss": 1.808, "step": 617 }, { "epoch": 0.06678914946503836, "grad_norm": 0.33646243810653687, "learning_rate": 8e-05, "loss": 1.7959, "step": 618 }, { "epoch": 0.06689722252242516, "grad_norm": 0.34386146068573, "learning_rate": 8e-05, "loss": 1.8981, "step": 619 }, { "epoch": 0.06700529557981195, "grad_norm": 0.378597617149353, "learning_rate": 8e-05, "loss": 1.7251, "step": 620 }, { "epoch": 0.06711336863719874, "grad_norm": 0.3454154133796692, "learning_rate": 8e-05, "loss": 1.6371, "step": 621 }, { "epoch": 0.06722144169458553, "grad_norm": 0.34818366169929504, "learning_rate": 8e-05, "loss": 1.7301, "step": 622 }, { "epoch": 0.06732951475197234, "grad_norm": 0.36195915937423706, "learning_rate": 8e-05, "loss": 1.9585, "step": 623 }, { "epoch": 0.06743758780935913, "grad_norm": 0.37999871373176575, "learning_rate": 8e-05, "loss": 1.8231, "step": 624 }, { "epoch": 0.06754566086674592, "grad_norm": 0.35395121574401855, "learning_rate": 8e-05, "loss": 1.8561, "step": 625 }, { "epoch": 0.06765373392413271, "grad_norm": 0.36281928420066833, "learning_rate": 8e-05, "loss": 1.9698, "step": 626 }, { "epoch": 0.06776180698151951, "grad_norm": 0.39710533618927, "learning_rate": 8e-05, "loss": 1.9219, "step": 627 }, { "epoch": 0.0678698800389063, "grad_norm": 0.3393266797065735, "learning_rate": 8e-05, "loss": 1.7098, "step": 628 }, { "epoch": 0.06797795309629309, "grad_norm": 0.368127703666687, "learning_rate": 8e-05, "loss": 1.7767, "step": 629 }, { "epoch": 0.06808602615367988, "grad_norm": 0.3395160734653473, "learning_rate": 8e-05, "loss": 1.7817, "step": 630 }, { "epoch": 0.06819409921106669, "grad_norm": 0.3301742970943451, "learning_rate": 8e-05, "loss": 1.7479, "step": 631 }, { "epoch": 0.06830217226845348, "grad_norm": 0.40448659658432007, "learning_rate": 8e-05, "loss": 1.9478, "step": 632 }, { "epoch": 0.06841024532584027, "grad_norm": 0.3483118712902069, "learning_rate": 8e-05, "loss": 1.8395, "step": 633 }, { "epoch": 0.06851831838322706, "grad_norm": 0.3806949853897095, "learning_rate": 8e-05, "loss": 1.9625, "step": 634 }, { "epoch": 0.06862639144061386, "grad_norm": 0.3638564944267273, "learning_rate": 8e-05, "loss": 1.9365, "step": 635 }, { "epoch": 0.06873446449800065, "grad_norm": 0.34495770931243896, "learning_rate": 8e-05, "loss": 1.8124, "step": 636 }, { "epoch": 0.06884253755538744, "grad_norm": 0.40036216378211975, "learning_rate": 8e-05, "loss": 2.0827, "step": 637 }, { "epoch": 0.06895061061277423, "grad_norm": 0.32573381066322327, "learning_rate": 8e-05, "loss": 1.736, "step": 638 }, { "epoch": 0.06905868367016103, "grad_norm": 0.34161123633384705, "learning_rate": 8e-05, "loss": 1.9109, "step": 639 }, { "epoch": 0.06916675672754782, "grad_norm": 0.35865241289138794, "learning_rate": 8e-05, "loss": 1.8919, "step": 640 }, { "epoch": 0.06927482978493461, "grad_norm": 0.3654935657978058, "learning_rate": 8e-05, "loss": 1.6172, "step": 641 }, { "epoch": 0.0693829028423214, "grad_norm": 0.32270491123199463, "learning_rate": 8e-05, "loss": 1.8346, "step": 642 }, { "epoch": 0.06949097589970821, "grad_norm": 0.3278350830078125, "learning_rate": 8e-05, "loss": 1.6747, "step": 643 }, { "epoch": 0.069599048957095, "grad_norm": 0.3394119441509247, "learning_rate": 8e-05, "loss": 1.8761, "step": 644 }, { "epoch": 0.06970712201448179, "grad_norm": 0.3506968319416046, "learning_rate": 8e-05, "loss": 1.7395, "step": 645 }, { "epoch": 0.06981519507186858, "grad_norm": 0.4216964542865753, "learning_rate": 8e-05, "loss": 1.9973, "step": 646 }, { "epoch": 0.06992326812925538, "grad_norm": 0.32112473249435425, "learning_rate": 8e-05, "loss": 1.6835, "step": 647 }, { "epoch": 0.07003134118664217, "grad_norm": 0.3561711311340332, "learning_rate": 8e-05, "loss": 1.6375, "step": 648 }, { "epoch": 0.07013941424402896, "grad_norm": 0.3361760675907135, "learning_rate": 8e-05, "loss": 1.8433, "step": 649 }, { "epoch": 0.07024748730141575, "grad_norm": 0.35475993156433105, "learning_rate": 8e-05, "loss": 1.7712, "step": 650 }, { "epoch": 0.07035556035880255, "grad_norm": 0.3732445538043976, "learning_rate": 8e-05, "loss": 1.729, "step": 651 }, { "epoch": 0.07046363341618934, "grad_norm": 0.36751070618629456, "learning_rate": 8e-05, "loss": 1.9178, "step": 652 }, { "epoch": 0.07057170647357613, "grad_norm": 0.3510911762714386, "learning_rate": 8e-05, "loss": 1.7829, "step": 653 }, { "epoch": 0.07067977953096292, "grad_norm": 0.3700132966041565, "learning_rate": 8e-05, "loss": 1.8732, "step": 654 }, { "epoch": 0.07078785258834973, "grad_norm": 0.3228524923324585, "learning_rate": 8e-05, "loss": 1.7292, "step": 655 }, { "epoch": 0.07089592564573652, "grad_norm": 0.3509063720703125, "learning_rate": 8e-05, "loss": 1.8732, "step": 656 }, { "epoch": 0.07100399870312331, "grad_norm": 0.3533901274204254, "learning_rate": 8e-05, "loss": 1.7774, "step": 657 }, { "epoch": 0.0711120717605101, "grad_norm": 0.34663504362106323, "learning_rate": 8e-05, "loss": 1.876, "step": 658 }, { "epoch": 0.0712201448178969, "grad_norm": 0.3430791199207306, "learning_rate": 8e-05, "loss": 1.7164, "step": 659 }, { "epoch": 0.07132821787528369, "grad_norm": 0.33836594223976135, "learning_rate": 8e-05, "loss": 1.824, "step": 660 }, { "epoch": 0.07143629093267048, "grad_norm": 0.3621346950531006, "learning_rate": 8e-05, "loss": 1.8906, "step": 661 }, { "epoch": 0.07154436399005727, "grad_norm": 0.34158116579055786, "learning_rate": 8e-05, "loss": 1.7485, "step": 662 }, { "epoch": 0.07165243704744408, "grad_norm": 0.37157899141311646, "learning_rate": 8e-05, "loss": 1.7956, "step": 663 }, { "epoch": 0.07176051010483087, "grad_norm": 0.355392187833786, "learning_rate": 8e-05, "loss": 1.8269, "step": 664 }, { "epoch": 0.07186858316221766, "grad_norm": 0.3327278196811676, "learning_rate": 8e-05, "loss": 1.8037, "step": 665 }, { "epoch": 0.07197665621960445, "grad_norm": 0.33161213994026184, "learning_rate": 8e-05, "loss": 1.7282, "step": 666 }, { "epoch": 0.07208472927699125, "grad_norm": 0.3450239598751068, "learning_rate": 8e-05, "loss": 1.677, "step": 667 }, { "epoch": 0.07219280233437804, "grad_norm": 0.3507198691368103, "learning_rate": 8e-05, "loss": 1.6889, "step": 668 }, { "epoch": 0.07230087539176483, "grad_norm": 0.34202083945274353, "learning_rate": 8e-05, "loss": 1.7022, "step": 669 }, { "epoch": 0.07240894844915163, "grad_norm": 0.3287566006183624, "learning_rate": 8e-05, "loss": 1.7038, "step": 670 }, { "epoch": 0.07251702150653842, "grad_norm": 0.36855319142341614, "learning_rate": 8e-05, "loss": 1.8515, "step": 671 }, { "epoch": 0.07262509456392521, "grad_norm": 0.340627521276474, "learning_rate": 8e-05, "loss": 1.8783, "step": 672 }, { "epoch": 0.072733167621312, "grad_norm": 0.3321596086025238, "learning_rate": 8e-05, "loss": 1.7327, "step": 673 }, { "epoch": 0.07284124067869881, "grad_norm": 0.3582914471626282, "learning_rate": 8e-05, "loss": 1.7415, "step": 674 }, { "epoch": 0.0729493137360856, "grad_norm": 0.3498310446739197, "learning_rate": 8e-05, "loss": 1.8232, "step": 675 }, { "epoch": 0.07305738679347239, "grad_norm": 0.31563234329223633, "learning_rate": 8e-05, "loss": 1.5993, "step": 676 }, { "epoch": 0.07316545985085918, "grad_norm": 0.3766508400440216, "learning_rate": 8e-05, "loss": 1.9739, "step": 677 }, { "epoch": 0.07327353290824598, "grad_norm": 0.34974056482315063, "learning_rate": 8e-05, "loss": 1.7983, "step": 678 }, { "epoch": 0.07338160596563277, "grad_norm": 0.3550325930118561, "learning_rate": 8e-05, "loss": 1.8809, "step": 679 }, { "epoch": 0.07348967902301956, "grad_norm": 0.4403340518474579, "learning_rate": 8e-05, "loss": 2.1644, "step": 680 }, { "epoch": 0.07359775208040635, "grad_norm": 0.3214254379272461, "learning_rate": 8e-05, "loss": 1.6729, "step": 681 }, { "epoch": 0.07370582513779315, "grad_norm": 0.3804623484611511, "learning_rate": 8e-05, "loss": 1.9349, "step": 682 }, { "epoch": 0.07381389819517994, "grad_norm": 0.32864585518836975, "learning_rate": 8e-05, "loss": 1.7857, "step": 683 }, { "epoch": 0.07392197125256673, "grad_norm": 0.35231640934944153, "learning_rate": 8e-05, "loss": 1.6804, "step": 684 }, { "epoch": 0.07403004430995352, "grad_norm": 0.3733230233192444, "learning_rate": 8e-05, "loss": 1.8922, "step": 685 }, { "epoch": 0.07413811736734033, "grad_norm": 0.3392426669597626, "learning_rate": 8e-05, "loss": 1.7012, "step": 686 }, { "epoch": 0.07424619042472712, "grad_norm": 0.3542892336845398, "learning_rate": 8e-05, "loss": 1.5467, "step": 687 }, { "epoch": 0.07435426348211391, "grad_norm": 0.35224273800849915, "learning_rate": 8e-05, "loss": 1.7769, "step": 688 }, { "epoch": 0.0744623365395007, "grad_norm": 0.3452960252761841, "learning_rate": 8e-05, "loss": 1.6577, "step": 689 }, { "epoch": 0.0745704095968875, "grad_norm": 0.3708077073097229, "learning_rate": 8e-05, "loss": 1.8156, "step": 690 }, { "epoch": 0.07467848265427429, "grad_norm": 0.3276267349720001, "learning_rate": 8e-05, "loss": 1.7198, "step": 691 }, { "epoch": 0.07478655571166108, "grad_norm": 0.3788032829761505, "learning_rate": 8e-05, "loss": 1.8047, "step": 692 }, { "epoch": 0.07489462876904787, "grad_norm": 0.37438222765922546, "learning_rate": 8e-05, "loss": 1.7688, "step": 693 }, { "epoch": 0.07500270182643468, "grad_norm": 0.37413498759269714, "learning_rate": 8e-05, "loss": 1.7814, "step": 694 }, { "epoch": 0.07511077488382147, "grad_norm": 0.33975186944007874, "learning_rate": 8e-05, "loss": 1.7363, "step": 695 }, { "epoch": 0.07521884794120826, "grad_norm": 0.34187236428260803, "learning_rate": 8e-05, "loss": 1.7086, "step": 696 }, { "epoch": 0.07532692099859505, "grad_norm": 0.3947480320930481, "learning_rate": 8e-05, "loss": 1.7987, "step": 697 }, { "epoch": 0.07543499405598185, "grad_norm": 0.44009360671043396, "learning_rate": 8e-05, "loss": 2.0018, "step": 698 }, { "epoch": 0.07554306711336864, "grad_norm": 0.33826857805252075, "learning_rate": 8e-05, "loss": 1.6599, "step": 699 }, { "epoch": 0.07565114017075543, "grad_norm": 0.3385428488254547, "learning_rate": 8e-05, "loss": 1.563, "step": 700 }, { "epoch": 0.07575921322814222, "grad_norm": 0.43305906653404236, "learning_rate": 8e-05, "loss": 1.9682, "step": 701 }, { "epoch": 0.07586728628552902, "grad_norm": 0.3494119644165039, "learning_rate": 8e-05, "loss": 1.7362, "step": 702 }, { "epoch": 0.07597535934291581, "grad_norm": 0.38518205285072327, "learning_rate": 8e-05, "loss": 1.9468, "step": 703 }, { "epoch": 0.0760834324003026, "grad_norm": 0.3591043949127197, "learning_rate": 8e-05, "loss": 1.8352, "step": 704 }, { "epoch": 0.07619150545768939, "grad_norm": 0.34808701276779175, "learning_rate": 8e-05, "loss": 1.8679, "step": 705 }, { "epoch": 0.0762995785150762, "grad_norm": 0.34616905450820923, "learning_rate": 8e-05, "loss": 1.8232, "step": 706 }, { "epoch": 0.07640765157246299, "grad_norm": 0.3296198546886444, "learning_rate": 8e-05, "loss": 1.8433, "step": 707 }, { "epoch": 0.07651572462984978, "grad_norm": 0.36884406208992004, "learning_rate": 8e-05, "loss": 1.8098, "step": 708 }, { "epoch": 0.07662379768723657, "grad_norm": 0.3298288881778717, "learning_rate": 8e-05, "loss": 1.6976, "step": 709 }, { "epoch": 0.07673187074462337, "grad_norm": 0.32953813672065735, "learning_rate": 8e-05, "loss": 1.625, "step": 710 }, { "epoch": 0.07683994380201016, "grad_norm": 0.35480138659477234, "learning_rate": 8e-05, "loss": 1.7083, "step": 711 }, { "epoch": 0.07694801685939695, "grad_norm": 0.3599317967891693, "learning_rate": 8e-05, "loss": 1.84, "step": 712 }, { "epoch": 0.07705608991678374, "grad_norm": 0.35181501507759094, "learning_rate": 8e-05, "loss": 1.7176, "step": 713 }, { "epoch": 0.07716416297417054, "grad_norm": 0.3353733420372009, "learning_rate": 8e-05, "loss": 1.8438, "step": 714 }, { "epoch": 0.07727223603155733, "grad_norm": 0.35124659538269043, "learning_rate": 8e-05, "loss": 1.938, "step": 715 }, { "epoch": 0.07738030908894412, "grad_norm": 0.4020790457725525, "learning_rate": 8e-05, "loss": 1.8665, "step": 716 }, { "epoch": 0.07748838214633091, "grad_norm": 0.3336130380630493, "learning_rate": 8e-05, "loss": 1.6259, "step": 717 }, { "epoch": 0.07759645520371772, "grad_norm": 0.32654890418052673, "learning_rate": 8e-05, "loss": 1.6961, "step": 718 }, { "epoch": 0.07770452826110451, "grad_norm": 0.35890254378318787, "learning_rate": 8e-05, "loss": 1.8516, "step": 719 }, { "epoch": 0.0778126013184913, "grad_norm": 0.3621284067630768, "learning_rate": 8e-05, "loss": 1.8362, "step": 720 }, { "epoch": 0.07792067437587809, "grad_norm": 0.36673128604888916, "learning_rate": 8e-05, "loss": 1.8275, "step": 721 }, { "epoch": 0.07802874743326489, "grad_norm": 0.34916919469833374, "learning_rate": 8e-05, "loss": 1.6601, "step": 722 }, { "epoch": 0.07813682049065168, "grad_norm": 0.33866086602211, "learning_rate": 8e-05, "loss": 1.6307, "step": 723 }, { "epoch": 0.07824489354803847, "grad_norm": 0.36578813195228577, "learning_rate": 8e-05, "loss": 1.8078, "step": 724 }, { "epoch": 0.07835296660542526, "grad_norm": 0.3880551755428314, "learning_rate": 8e-05, "loss": 1.8684, "step": 725 }, { "epoch": 0.07846103966281207, "grad_norm": 0.3700942099094391, "learning_rate": 8e-05, "loss": 1.909, "step": 726 }, { "epoch": 0.07856911272019886, "grad_norm": 0.4181189239025116, "learning_rate": 8e-05, "loss": 1.8998, "step": 727 }, { "epoch": 0.07867718577758565, "grad_norm": 0.4106231927871704, "learning_rate": 8e-05, "loss": 1.7916, "step": 728 }, { "epoch": 0.07878525883497244, "grad_norm": 0.403994083404541, "learning_rate": 8e-05, "loss": 1.8129, "step": 729 }, { "epoch": 0.07889333189235924, "grad_norm": 0.3424227833747864, "learning_rate": 8e-05, "loss": 1.722, "step": 730 }, { "epoch": 0.07900140494974603, "grad_norm": 0.3473060727119446, "learning_rate": 8e-05, "loss": 1.8429, "step": 731 }, { "epoch": 0.07910947800713282, "grad_norm": 0.37217941880226135, "learning_rate": 8e-05, "loss": 1.8698, "step": 732 }, { "epoch": 0.07921755106451961, "grad_norm": 0.3986811637878418, "learning_rate": 8e-05, "loss": 1.8949, "step": 733 }, { "epoch": 0.07932562412190641, "grad_norm": 0.382118284702301, "learning_rate": 8e-05, "loss": 1.9121, "step": 734 }, { "epoch": 0.0794336971792932, "grad_norm": 0.34268704056739807, "learning_rate": 8e-05, "loss": 1.7358, "step": 735 }, { "epoch": 0.07954177023667999, "grad_norm": 0.3592139482498169, "learning_rate": 8e-05, "loss": 1.9336, "step": 736 }, { "epoch": 0.07964984329406678, "grad_norm": 0.36988234519958496, "learning_rate": 8e-05, "loss": 1.8503, "step": 737 }, { "epoch": 0.07975791635145359, "grad_norm": 0.38400474190711975, "learning_rate": 8e-05, "loss": 1.9291, "step": 738 }, { "epoch": 0.07986598940884038, "grad_norm": 0.3458898961544037, "learning_rate": 8e-05, "loss": 1.7461, "step": 739 }, { "epoch": 0.07997406246622717, "grad_norm": 0.3562629818916321, "learning_rate": 8e-05, "loss": 1.8328, "step": 740 }, { "epoch": 0.08008213552361396, "grad_norm": 0.36015400290489197, "learning_rate": 8e-05, "loss": 1.7099, "step": 741 }, { "epoch": 0.08019020858100076, "grad_norm": 0.3556027114391327, "learning_rate": 8e-05, "loss": 1.7381, "step": 742 }, { "epoch": 0.08029828163838755, "grad_norm": 0.3555619716644287, "learning_rate": 8e-05, "loss": 1.8762, "step": 743 }, { "epoch": 0.08040635469577434, "grad_norm": 0.3548396825790405, "learning_rate": 8e-05, "loss": 1.895, "step": 744 }, { "epoch": 0.08051442775316113, "grad_norm": 0.352502703666687, "learning_rate": 8e-05, "loss": 1.8113, "step": 745 }, { "epoch": 0.08062250081054793, "grad_norm": 0.372275173664093, "learning_rate": 8e-05, "loss": 1.7969, "step": 746 }, { "epoch": 0.08073057386793472, "grad_norm": 0.3499962389469147, "learning_rate": 8e-05, "loss": 1.8736, "step": 747 }, { "epoch": 0.08083864692532151, "grad_norm": 0.35915976762771606, "learning_rate": 8e-05, "loss": 1.8952, "step": 748 }, { "epoch": 0.0809467199827083, "grad_norm": 0.39708012342453003, "learning_rate": 8e-05, "loss": 2.0697, "step": 749 }, { "epoch": 0.08105479304009511, "grad_norm": 0.3751751184463501, "learning_rate": 8e-05, "loss": 1.8342, "step": 750 }, { "epoch": 0.0811628660974819, "grad_norm": 0.37397319078445435, "learning_rate": 8e-05, "loss": 1.7358, "step": 751 }, { "epoch": 0.08127093915486869, "grad_norm": 0.3618088960647583, "learning_rate": 8e-05, "loss": 1.7554, "step": 752 }, { "epoch": 0.08137901221225549, "grad_norm": 0.3574982285499573, "learning_rate": 8e-05, "loss": 1.8672, "step": 753 }, { "epoch": 0.08148708526964228, "grad_norm": 0.386802613735199, "learning_rate": 8e-05, "loss": 1.9458, "step": 754 }, { "epoch": 0.08159515832702907, "grad_norm": 0.36350077390670776, "learning_rate": 8e-05, "loss": 1.9469, "step": 755 }, { "epoch": 0.08170323138441586, "grad_norm": 0.33847320079803467, "learning_rate": 8e-05, "loss": 1.7246, "step": 756 }, { "epoch": 0.08181130444180267, "grad_norm": 0.3289634585380554, "learning_rate": 8e-05, "loss": 1.7402, "step": 757 }, { "epoch": 0.08191937749918946, "grad_norm": 0.32910841703414917, "learning_rate": 8e-05, "loss": 1.6152, "step": 758 }, { "epoch": 0.08202745055657625, "grad_norm": 0.3602372407913208, "learning_rate": 8e-05, "loss": 1.8185, "step": 759 }, { "epoch": 0.08213552361396304, "grad_norm": 0.3620285987854004, "learning_rate": 8e-05, "loss": 1.928, "step": 760 }, { "epoch": 0.08224359667134984, "grad_norm": 0.3445507884025574, "learning_rate": 8e-05, "loss": 1.9045, "step": 761 }, { "epoch": 0.08235166972873663, "grad_norm": 0.3363454043865204, "learning_rate": 8e-05, "loss": 1.7501, "step": 762 }, { "epoch": 0.08245974278612342, "grad_norm": 0.3430785536766052, "learning_rate": 8e-05, "loss": 1.5866, "step": 763 }, { "epoch": 0.08256781584351021, "grad_norm": 0.3677377700805664, "learning_rate": 8e-05, "loss": 2.0467, "step": 764 }, { "epoch": 0.08267588890089701, "grad_norm": 0.35696345567703247, "learning_rate": 8e-05, "loss": 1.824, "step": 765 }, { "epoch": 0.0827839619582838, "grad_norm": 0.3691982924938202, "learning_rate": 8e-05, "loss": 1.9481, "step": 766 }, { "epoch": 0.08289203501567059, "grad_norm": 0.33297982811927795, "learning_rate": 8e-05, "loss": 1.634, "step": 767 }, { "epoch": 0.08300010807305738, "grad_norm": 0.3267819881439209, "learning_rate": 8e-05, "loss": 1.6767, "step": 768 }, { "epoch": 0.08310818113044419, "grad_norm": 0.3792954385280609, "learning_rate": 8e-05, "loss": 1.7697, "step": 769 }, { "epoch": 0.08321625418783098, "grad_norm": 0.3626005947589874, "learning_rate": 8e-05, "loss": 1.8389, "step": 770 }, { "epoch": 0.08332432724521777, "grad_norm": 0.37124645709991455, "learning_rate": 8e-05, "loss": 1.821, "step": 771 }, { "epoch": 0.08343240030260456, "grad_norm": 0.3513961732387543, "learning_rate": 8e-05, "loss": 1.7428, "step": 772 }, { "epoch": 0.08354047335999136, "grad_norm": 0.34473925828933716, "learning_rate": 8e-05, "loss": 1.7266, "step": 773 }, { "epoch": 0.08364854641737815, "grad_norm": 0.34735044836997986, "learning_rate": 8e-05, "loss": 1.7383, "step": 774 }, { "epoch": 0.08375661947476494, "grad_norm": 0.37575075030326843, "learning_rate": 8e-05, "loss": 1.6575, "step": 775 }, { "epoch": 0.08386469253215173, "grad_norm": 0.39621371030807495, "learning_rate": 8e-05, "loss": 1.9143, "step": 776 }, { "epoch": 0.08397276558953853, "grad_norm": 0.3341212570667267, "learning_rate": 8e-05, "loss": 1.7078, "step": 777 }, { "epoch": 0.08408083864692532, "grad_norm": 0.4078580141067505, "learning_rate": 8e-05, "loss": 1.8777, "step": 778 }, { "epoch": 0.08418891170431211, "grad_norm": 0.3987581431865692, "learning_rate": 8e-05, "loss": 1.923, "step": 779 }, { "epoch": 0.0842969847616989, "grad_norm": 0.3557502031326294, "learning_rate": 8e-05, "loss": 1.6162, "step": 780 }, { "epoch": 0.08440505781908571, "grad_norm": 0.32769715785980225, "learning_rate": 8e-05, "loss": 1.5147, "step": 781 }, { "epoch": 0.0845131308764725, "grad_norm": 0.35191041231155396, "learning_rate": 8e-05, "loss": 1.6752, "step": 782 }, { "epoch": 0.08462120393385929, "grad_norm": 0.3392881155014038, "learning_rate": 8e-05, "loss": 1.6434, "step": 783 }, { "epoch": 0.08472927699124608, "grad_norm": 0.33011457324028015, "learning_rate": 8e-05, "loss": 1.7721, "step": 784 }, { "epoch": 0.08483735004863288, "grad_norm": 0.3555687963962555, "learning_rate": 8e-05, "loss": 1.7182, "step": 785 }, { "epoch": 0.08494542310601967, "grad_norm": 0.3733087480068207, "learning_rate": 8e-05, "loss": 2.0021, "step": 786 }, { "epoch": 0.08505349616340646, "grad_norm": 0.36685800552368164, "learning_rate": 8e-05, "loss": 1.8644, "step": 787 }, { "epoch": 0.08516156922079325, "grad_norm": 0.3459504246711731, "learning_rate": 8e-05, "loss": 1.7951, "step": 788 }, { "epoch": 0.08526964227818006, "grad_norm": 0.36492791771888733, "learning_rate": 8e-05, "loss": 1.8778, "step": 789 }, { "epoch": 0.08537771533556685, "grad_norm": 0.3297025263309479, "learning_rate": 8e-05, "loss": 1.7434, "step": 790 }, { "epoch": 0.08548578839295363, "grad_norm": 0.32382282614707947, "learning_rate": 8e-05, "loss": 1.7046, "step": 791 }, { "epoch": 0.08559386145034042, "grad_norm": 0.36852508783340454, "learning_rate": 8e-05, "loss": 1.8681, "step": 792 }, { "epoch": 0.08570193450772723, "grad_norm": 0.3332373797893524, "learning_rate": 8e-05, "loss": 1.7354, "step": 793 }, { "epoch": 0.08581000756511402, "grad_norm": 0.32317543029785156, "learning_rate": 8e-05, "loss": 1.7088, "step": 794 }, { "epoch": 0.08591808062250081, "grad_norm": 0.34304043650627136, "learning_rate": 8e-05, "loss": 1.7349, "step": 795 }, { "epoch": 0.0860261536798876, "grad_norm": 0.35308459401130676, "learning_rate": 8e-05, "loss": 1.9224, "step": 796 }, { "epoch": 0.0861342267372744, "grad_norm": 0.37210604548454285, "learning_rate": 8e-05, "loss": 1.9415, "step": 797 }, { "epoch": 0.08624229979466119, "grad_norm": 0.336307168006897, "learning_rate": 8e-05, "loss": 1.6663, "step": 798 }, { "epoch": 0.08635037285204798, "grad_norm": 0.3659408986568451, "learning_rate": 8e-05, "loss": 1.7261, "step": 799 }, { "epoch": 0.08645844590943477, "grad_norm": 0.334870308637619, "learning_rate": 8e-05, "loss": 1.6286, "step": 800 }, { "epoch": 0.08656651896682158, "grad_norm": 0.37102147936820984, "learning_rate": 8e-05, "loss": 1.9711, "step": 801 }, { "epoch": 0.08667459202420837, "grad_norm": 0.37157055735588074, "learning_rate": 8e-05, "loss": 1.6847, "step": 802 }, { "epoch": 0.08678266508159516, "grad_norm": 0.34305718541145325, "learning_rate": 8e-05, "loss": 1.7147, "step": 803 }, { "epoch": 0.08689073813898195, "grad_norm": 0.3572945296764374, "learning_rate": 8e-05, "loss": 1.9602, "step": 804 }, { "epoch": 0.08699881119636875, "grad_norm": 0.36801835894584656, "learning_rate": 8e-05, "loss": 1.648, "step": 805 }, { "epoch": 0.08710688425375554, "grad_norm": 0.3747352361679077, "learning_rate": 8e-05, "loss": 1.9976, "step": 806 }, { "epoch": 0.08721495731114233, "grad_norm": 0.37714725732803345, "learning_rate": 8e-05, "loss": 1.5962, "step": 807 }, { "epoch": 0.08732303036852912, "grad_norm": 0.36145877838134766, "learning_rate": 8e-05, "loss": 1.8676, "step": 808 }, { "epoch": 0.08743110342591592, "grad_norm": 0.33727219700813293, "learning_rate": 8e-05, "loss": 1.8809, "step": 809 }, { "epoch": 0.08753917648330271, "grad_norm": 0.4171218276023865, "learning_rate": 8e-05, "loss": 1.929, "step": 810 }, { "epoch": 0.0876472495406895, "grad_norm": 0.3338264226913452, "learning_rate": 8e-05, "loss": 1.7619, "step": 811 }, { "epoch": 0.0877553225980763, "grad_norm": 0.32613831758499146, "learning_rate": 8e-05, "loss": 1.6547, "step": 812 }, { "epoch": 0.0878633956554631, "grad_norm": 0.33511805534362793, "learning_rate": 8e-05, "loss": 1.7767, "step": 813 }, { "epoch": 0.08797146871284989, "grad_norm": 0.33658602833747864, "learning_rate": 8e-05, "loss": 1.8978, "step": 814 }, { "epoch": 0.08807954177023668, "grad_norm": 0.3420501947402954, "learning_rate": 8e-05, "loss": 1.6778, "step": 815 }, { "epoch": 0.08818761482762347, "grad_norm": 0.336590975522995, "learning_rate": 8e-05, "loss": 1.7688, "step": 816 }, { "epoch": 0.08829568788501027, "grad_norm": 0.36553964018821716, "learning_rate": 8e-05, "loss": 1.7911, "step": 817 }, { "epoch": 0.08840376094239706, "grad_norm": 0.34292399883270264, "learning_rate": 8e-05, "loss": 1.8965, "step": 818 }, { "epoch": 0.08851183399978385, "grad_norm": 0.3505082130432129, "learning_rate": 8e-05, "loss": 1.8568, "step": 819 }, { "epoch": 0.08861990705717064, "grad_norm": 0.3620678782463074, "learning_rate": 8e-05, "loss": 1.7478, "step": 820 }, { "epoch": 0.08872798011455744, "grad_norm": 0.3877421021461487, "learning_rate": 8e-05, "loss": 1.9394, "step": 821 }, { "epoch": 0.08883605317194423, "grad_norm": 0.36055201292037964, "learning_rate": 8e-05, "loss": 1.6753, "step": 822 }, { "epoch": 0.08894412622933102, "grad_norm": 0.36477819085121155, "learning_rate": 8e-05, "loss": 1.9841, "step": 823 }, { "epoch": 0.08905219928671781, "grad_norm": 0.3890232741832733, "learning_rate": 8e-05, "loss": 1.5518, "step": 824 }, { "epoch": 0.08916027234410462, "grad_norm": 0.3630181849002838, "learning_rate": 8e-05, "loss": 1.8396, "step": 825 }, { "epoch": 0.08926834540149141, "grad_norm": 0.3804430663585663, "learning_rate": 8e-05, "loss": 1.994, "step": 826 }, { "epoch": 0.0893764184588782, "grad_norm": 0.33240169286727905, "learning_rate": 8e-05, "loss": 1.7174, "step": 827 }, { "epoch": 0.08948449151626499, "grad_norm": 0.3295738399028778, "learning_rate": 8e-05, "loss": 1.6854, "step": 828 }, { "epoch": 0.08959256457365179, "grad_norm": 0.34137552976608276, "learning_rate": 8e-05, "loss": 1.7784, "step": 829 }, { "epoch": 0.08970063763103858, "grad_norm": 0.35221177339553833, "learning_rate": 8e-05, "loss": 1.8779, "step": 830 }, { "epoch": 0.08980871068842537, "grad_norm": 0.36139872670173645, "learning_rate": 8e-05, "loss": 1.797, "step": 831 }, { "epoch": 0.08991678374581216, "grad_norm": 0.3533114194869995, "learning_rate": 8e-05, "loss": 1.7841, "step": 832 }, { "epoch": 0.09002485680319897, "grad_norm": 0.3557817041873932, "learning_rate": 8e-05, "loss": 1.7212, "step": 833 }, { "epoch": 0.09013292986058576, "grad_norm": 0.32702866196632385, "learning_rate": 8e-05, "loss": 1.6826, "step": 834 }, { "epoch": 0.09024100291797255, "grad_norm": 0.33852875232696533, "learning_rate": 8e-05, "loss": 1.7097, "step": 835 }, { "epoch": 0.09034907597535935, "grad_norm": 0.3507237732410431, "learning_rate": 8e-05, "loss": 1.7899, "step": 836 }, { "epoch": 0.09045714903274614, "grad_norm": 0.39895108342170715, "learning_rate": 8e-05, "loss": 1.9812, "step": 837 }, { "epoch": 0.09056522209013293, "grad_norm": 0.3445395231246948, "learning_rate": 8e-05, "loss": 1.4396, "step": 838 }, { "epoch": 0.09067329514751972, "grad_norm": 0.3465981185436249, "learning_rate": 8e-05, "loss": 1.7677, "step": 839 }, { "epoch": 0.09078136820490652, "grad_norm": 0.3418671786785126, "learning_rate": 8e-05, "loss": 1.7083, "step": 840 }, { "epoch": 0.09088944126229331, "grad_norm": 0.3876403570175171, "learning_rate": 8e-05, "loss": 2.0084, "step": 841 }, { "epoch": 0.0909975143196801, "grad_norm": 0.3500586152076721, "learning_rate": 8e-05, "loss": 1.9042, "step": 842 }, { "epoch": 0.0911055873770669, "grad_norm": 0.3533492982387543, "learning_rate": 8e-05, "loss": 1.7577, "step": 843 }, { "epoch": 0.0912136604344537, "grad_norm": 0.34509995579719543, "learning_rate": 8e-05, "loss": 1.8736, "step": 844 }, { "epoch": 0.09132173349184049, "grad_norm": 0.34286263585090637, "learning_rate": 8e-05, "loss": 1.7923, "step": 845 }, { "epoch": 0.09142980654922728, "grad_norm": 0.34014856815338135, "learning_rate": 8e-05, "loss": 1.7946, "step": 846 }, { "epoch": 0.09153787960661407, "grad_norm": 0.3614908754825592, "learning_rate": 8e-05, "loss": 1.7894, "step": 847 }, { "epoch": 0.09164595266400087, "grad_norm": 0.3748004138469696, "learning_rate": 8e-05, "loss": 1.8593, "step": 848 }, { "epoch": 0.09175402572138766, "grad_norm": 0.3322995603084564, "learning_rate": 8e-05, "loss": 1.5334, "step": 849 }, { "epoch": 0.09186209877877445, "grad_norm": 0.36351263523101807, "learning_rate": 8e-05, "loss": 1.8635, "step": 850 }, { "epoch": 0.09197017183616124, "grad_norm": 0.34182462096214294, "learning_rate": 8e-05, "loss": 1.7542, "step": 851 }, { "epoch": 0.09207824489354804, "grad_norm": 0.35077911615371704, "learning_rate": 8e-05, "loss": 1.9268, "step": 852 }, { "epoch": 0.09218631795093483, "grad_norm": 0.3587110638618469, "learning_rate": 8e-05, "loss": 1.64, "step": 853 }, { "epoch": 0.09229439100832162, "grad_norm": 0.3431946635246277, "learning_rate": 8e-05, "loss": 1.69, "step": 854 }, { "epoch": 0.09240246406570841, "grad_norm": 0.3534054458141327, "learning_rate": 8e-05, "loss": 1.7812, "step": 855 }, { "epoch": 0.09251053712309522, "grad_norm": 0.33270472288131714, "learning_rate": 8e-05, "loss": 1.8176, "step": 856 }, { "epoch": 0.09261861018048201, "grad_norm": 0.3302570879459381, "learning_rate": 8e-05, "loss": 1.6696, "step": 857 }, { "epoch": 0.0927266832378688, "grad_norm": 0.3459129333496094, "learning_rate": 8e-05, "loss": 1.7109, "step": 858 }, { "epoch": 0.09283475629525559, "grad_norm": 0.3526592552661896, "learning_rate": 8e-05, "loss": 1.8205, "step": 859 }, { "epoch": 0.09294282935264239, "grad_norm": 0.3434501886367798, "learning_rate": 8e-05, "loss": 1.5596, "step": 860 }, { "epoch": 0.09305090241002918, "grad_norm": 0.38379454612731934, "learning_rate": 8e-05, "loss": 1.8075, "step": 861 }, { "epoch": 0.09315897546741597, "grad_norm": 0.3195924758911133, "learning_rate": 8e-05, "loss": 1.6406, "step": 862 }, { "epoch": 0.09326704852480276, "grad_norm": 0.3645499348640442, "learning_rate": 8e-05, "loss": 1.7899, "step": 863 }, { "epoch": 0.09337512158218957, "grad_norm": 0.3458186686038971, "learning_rate": 8e-05, "loss": 1.6969, "step": 864 }, { "epoch": 0.09348319463957636, "grad_norm": 0.3764459490776062, "learning_rate": 8e-05, "loss": 1.8335, "step": 865 }, { "epoch": 0.09359126769696315, "grad_norm": 0.3541610538959503, "learning_rate": 8e-05, "loss": 1.63, "step": 866 }, { "epoch": 0.09369934075434994, "grad_norm": 0.3427675664424896, "learning_rate": 8e-05, "loss": 1.8361, "step": 867 }, { "epoch": 0.09380741381173674, "grad_norm": 0.37080663442611694, "learning_rate": 8e-05, "loss": 1.7066, "step": 868 }, { "epoch": 0.09391548686912353, "grad_norm": 0.3579069674015045, "learning_rate": 8e-05, "loss": 1.633, "step": 869 }, { "epoch": 0.09402355992651032, "grad_norm": 0.37710386514663696, "learning_rate": 8e-05, "loss": 1.7893, "step": 870 }, { "epoch": 0.09413163298389711, "grad_norm": 0.3282741606235504, "learning_rate": 8e-05, "loss": 1.6462, "step": 871 }, { "epoch": 0.09423970604128391, "grad_norm": 0.3715938627719879, "learning_rate": 8e-05, "loss": 1.8297, "step": 872 }, { "epoch": 0.0943477790986707, "grad_norm": 0.3862229585647583, "learning_rate": 8e-05, "loss": 1.7614, "step": 873 }, { "epoch": 0.0944558521560575, "grad_norm": 0.3499211072921753, "learning_rate": 8e-05, "loss": 1.7657, "step": 874 }, { "epoch": 0.09456392521344428, "grad_norm": 0.3432254493236542, "learning_rate": 8e-05, "loss": 1.8058, "step": 875 }, { "epoch": 0.09467199827083109, "grad_norm": 0.34992873668670654, "learning_rate": 8e-05, "loss": 1.7749, "step": 876 }, { "epoch": 0.09478007132821788, "grad_norm": 0.3644714653491974, "learning_rate": 8e-05, "loss": 1.8002, "step": 877 }, { "epoch": 0.09488814438560467, "grad_norm": 0.4058280289173126, "learning_rate": 8e-05, "loss": 1.7815, "step": 878 }, { "epoch": 0.09499621744299146, "grad_norm": 0.3541181683540344, "learning_rate": 8e-05, "loss": 1.8194, "step": 879 }, { "epoch": 0.09510429050037826, "grad_norm": 0.36470821499824524, "learning_rate": 8e-05, "loss": 1.9349, "step": 880 }, { "epoch": 0.09521236355776505, "grad_norm": 0.34866970777511597, "learning_rate": 8e-05, "loss": 1.8882, "step": 881 }, { "epoch": 0.09532043661515184, "grad_norm": 0.3558403551578522, "learning_rate": 8e-05, "loss": 1.941, "step": 882 }, { "epoch": 0.09542850967253863, "grad_norm": 0.3859046399593353, "learning_rate": 8e-05, "loss": 1.9231, "step": 883 }, { "epoch": 0.09553658272992543, "grad_norm": 0.36378538608551025, "learning_rate": 8e-05, "loss": 1.9172, "step": 884 }, { "epoch": 0.09564465578731222, "grad_norm": 0.3842751085758209, "learning_rate": 8e-05, "loss": 2.0204, "step": 885 }, { "epoch": 0.09575272884469901, "grad_norm": 0.37410861253738403, "learning_rate": 8e-05, "loss": 1.8148, "step": 886 }, { "epoch": 0.0958608019020858, "grad_norm": 0.38042762875556946, "learning_rate": 8e-05, "loss": 1.8696, "step": 887 }, { "epoch": 0.09596887495947261, "grad_norm": 0.3503395617008209, "learning_rate": 8e-05, "loss": 1.6475, "step": 888 }, { "epoch": 0.0960769480168594, "grad_norm": 0.3514329791069031, "learning_rate": 8e-05, "loss": 1.7627, "step": 889 }, { "epoch": 0.09618502107424619, "grad_norm": 0.3578362762928009, "learning_rate": 8e-05, "loss": 1.9048, "step": 890 }, { "epoch": 0.09629309413163298, "grad_norm": 0.38349369168281555, "learning_rate": 8e-05, "loss": 1.7972, "step": 891 }, { "epoch": 0.09640116718901978, "grad_norm": 0.39928188920021057, "learning_rate": 8e-05, "loss": 1.928, "step": 892 }, { "epoch": 0.09650924024640657, "grad_norm": 0.390762597322464, "learning_rate": 8e-05, "loss": 1.7685, "step": 893 }, { "epoch": 0.09661731330379336, "grad_norm": 0.34487417340278625, "learning_rate": 8e-05, "loss": 1.8621, "step": 894 }, { "epoch": 0.09672538636118015, "grad_norm": 0.3497505486011505, "learning_rate": 8e-05, "loss": 1.7434, "step": 895 }, { "epoch": 0.09683345941856696, "grad_norm": 0.42294076085090637, "learning_rate": 8e-05, "loss": 2.0171, "step": 896 }, { "epoch": 0.09694153247595375, "grad_norm": 0.36248984932899475, "learning_rate": 8e-05, "loss": 1.7623, "step": 897 }, { "epoch": 0.09704960553334054, "grad_norm": 0.35715553164482117, "learning_rate": 8e-05, "loss": 1.9705, "step": 898 }, { "epoch": 0.09715767859072733, "grad_norm": 0.34239330887794495, "learning_rate": 8e-05, "loss": 1.7118, "step": 899 }, { "epoch": 0.09726575164811413, "grad_norm": 0.3759861886501312, "learning_rate": 8e-05, "loss": 1.7913, "step": 900 }, { "epoch": 0.09737382470550092, "grad_norm": 0.40702810883522034, "learning_rate": 8e-05, "loss": 1.8557, "step": 901 }, { "epoch": 0.09748189776288771, "grad_norm": 0.32868754863739014, "learning_rate": 8e-05, "loss": 1.6483, "step": 902 }, { "epoch": 0.0975899708202745, "grad_norm": 0.35593172907829285, "learning_rate": 8e-05, "loss": 1.895, "step": 903 }, { "epoch": 0.0976980438776613, "grad_norm": 0.35074326395988464, "learning_rate": 8e-05, "loss": 1.7843, "step": 904 }, { "epoch": 0.09780611693504809, "grad_norm": 0.34221625328063965, "learning_rate": 8e-05, "loss": 1.7098, "step": 905 }, { "epoch": 0.09791418999243488, "grad_norm": 0.3473942279815674, "learning_rate": 8e-05, "loss": 1.7558, "step": 906 }, { "epoch": 0.09802226304982167, "grad_norm": 0.3827817440032959, "learning_rate": 8e-05, "loss": 1.7816, "step": 907 }, { "epoch": 0.09813033610720848, "grad_norm": 0.3702685236930847, "learning_rate": 8e-05, "loss": 1.9114, "step": 908 }, { "epoch": 0.09823840916459527, "grad_norm": 0.35614997148513794, "learning_rate": 8e-05, "loss": 1.8615, "step": 909 }, { "epoch": 0.09834648222198206, "grad_norm": 0.3270261287689209, "learning_rate": 8e-05, "loss": 1.7424, "step": 910 }, { "epoch": 0.09845455527936885, "grad_norm": 0.3622662425041199, "learning_rate": 8e-05, "loss": 1.7252, "step": 911 }, { "epoch": 0.09856262833675565, "grad_norm": 0.3668723404407501, "learning_rate": 8e-05, "loss": 1.8184, "step": 912 }, { "epoch": 0.09867070139414244, "grad_norm": 0.34703904390335083, "learning_rate": 8e-05, "loss": 1.7484, "step": 913 }, { "epoch": 0.09877877445152923, "grad_norm": 0.38205552101135254, "learning_rate": 8e-05, "loss": 1.8642, "step": 914 }, { "epoch": 0.09888684750891602, "grad_norm": 0.3451690971851349, "learning_rate": 8e-05, "loss": 1.6173, "step": 915 }, { "epoch": 0.09899492056630282, "grad_norm": 0.3449813723564148, "learning_rate": 8e-05, "loss": 1.7767, "step": 916 }, { "epoch": 0.09910299362368961, "grad_norm": 0.4120391607284546, "learning_rate": 8e-05, "loss": 2.0339, "step": 917 }, { "epoch": 0.0992110666810764, "grad_norm": 0.3527565598487854, "learning_rate": 8e-05, "loss": 1.8451, "step": 918 }, { "epoch": 0.09931913973846321, "grad_norm": 0.347257137298584, "learning_rate": 8e-05, "loss": 1.8731, "step": 919 }, { "epoch": 0.09942721279585, "grad_norm": 0.3716275990009308, "learning_rate": 8e-05, "loss": 1.9667, "step": 920 }, { "epoch": 0.09953528585323679, "grad_norm": 0.37038666009902954, "learning_rate": 8e-05, "loss": 1.8925, "step": 921 }, { "epoch": 0.09964335891062358, "grad_norm": 0.32822686433792114, "learning_rate": 8e-05, "loss": 1.6566, "step": 922 }, { "epoch": 0.09975143196801038, "grad_norm": 0.34583452343940735, "learning_rate": 8e-05, "loss": 1.7298, "step": 923 }, { "epoch": 0.09985950502539717, "grad_norm": 0.3826597332954407, "learning_rate": 8e-05, "loss": 1.8402, "step": 924 }, { "epoch": 0.09996757808278396, "grad_norm": 0.34853339195251465, "learning_rate": 8e-05, "loss": 1.6854, "step": 925 }, { "epoch": 0.10007565114017075, "grad_norm": 0.32940131425857544, "learning_rate": 8e-05, "loss": 1.6778, "step": 926 }, { "epoch": 0.10018372419755756, "grad_norm": 0.3581061065196991, "learning_rate": 8e-05, "loss": 1.7362, "step": 927 }, { "epoch": 0.10029179725494435, "grad_norm": 0.3408588469028473, "learning_rate": 8e-05, "loss": 1.7291, "step": 928 }, { "epoch": 0.10039987031233114, "grad_norm": 0.34937652945518494, "learning_rate": 8e-05, "loss": 1.7455, "step": 929 }, { "epoch": 0.10050794336971793, "grad_norm": 0.3451722264289856, "learning_rate": 8e-05, "loss": 1.5794, "step": 930 }, { "epoch": 0.10061601642710473, "grad_norm": 0.32624149322509766, "learning_rate": 8e-05, "loss": 1.7241, "step": 931 }, { "epoch": 0.10072408948449152, "grad_norm": 0.3210192322731018, "learning_rate": 8e-05, "loss": 1.5514, "step": 932 }, { "epoch": 0.10083216254187831, "grad_norm": 0.33881422877311707, "learning_rate": 8e-05, "loss": 1.7128, "step": 933 }, { "epoch": 0.1009402355992651, "grad_norm": 0.3912752866744995, "learning_rate": 8e-05, "loss": 1.852, "step": 934 }, { "epoch": 0.1010483086566519, "grad_norm": 0.36140796542167664, "learning_rate": 8e-05, "loss": 1.6899, "step": 935 }, { "epoch": 0.10115638171403869, "grad_norm": 0.3649536073207855, "learning_rate": 8e-05, "loss": 1.8636, "step": 936 }, { "epoch": 0.10126445477142548, "grad_norm": 0.3420322835445404, "learning_rate": 8e-05, "loss": 1.7951, "step": 937 }, { "epoch": 0.10137252782881227, "grad_norm": 0.36467358469963074, "learning_rate": 8e-05, "loss": 1.8276, "step": 938 }, { "epoch": 0.10148060088619908, "grad_norm": 0.3385829031467438, "learning_rate": 8e-05, "loss": 1.7406, "step": 939 }, { "epoch": 0.10158867394358587, "grad_norm": 0.34874477982521057, "learning_rate": 8e-05, "loss": 1.8694, "step": 940 }, { "epoch": 0.10169674700097266, "grad_norm": 0.35837405920028687, "learning_rate": 8e-05, "loss": 1.7741, "step": 941 }, { "epoch": 0.10180482005835945, "grad_norm": 0.40015703439712524, "learning_rate": 8e-05, "loss": 1.769, "step": 942 }, { "epoch": 0.10191289311574625, "grad_norm": 0.39776626229286194, "learning_rate": 8e-05, "loss": 2.0724, "step": 943 }, { "epoch": 0.10202096617313304, "grad_norm": 0.3498876392841339, "learning_rate": 8e-05, "loss": 1.8087, "step": 944 }, { "epoch": 0.10212903923051983, "grad_norm": 0.3590984046459198, "learning_rate": 8e-05, "loss": 1.7932, "step": 945 }, { "epoch": 0.10223711228790662, "grad_norm": 0.33495789766311646, "learning_rate": 8e-05, "loss": 1.7068, "step": 946 }, { "epoch": 0.10234518534529342, "grad_norm": 0.3117012083530426, "learning_rate": 8e-05, "loss": 1.5938, "step": 947 }, { "epoch": 0.10245325840268021, "grad_norm": 0.3461303412914276, "learning_rate": 8e-05, "loss": 1.6385, "step": 948 }, { "epoch": 0.102561331460067, "grad_norm": 0.36226168274879456, "learning_rate": 8e-05, "loss": 1.8516, "step": 949 }, { "epoch": 0.1026694045174538, "grad_norm": 0.3622657060623169, "learning_rate": 8e-05, "loss": 1.736, "step": 950 }, { "epoch": 0.1027774775748406, "grad_norm": 0.3543497920036316, "learning_rate": 8e-05, "loss": 1.7553, "step": 951 }, { "epoch": 0.10288555063222739, "grad_norm": 0.3250940144062042, "learning_rate": 8e-05, "loss": 1.7062, "step": 952 }, { "epoch": 0.10299362368961418, "grad_norm": 0.3531031310558319, "learning_rate": 8e-05, "loss": 1.8073, "step": 953 }, { "epoch": 0.10310169674700097, "grad_norm": 0.3336448669433594, "learning_rate": 8e-05, "loss": 1.7333, "step": 954 }, { "epoch": 0.10320976980438777, "grad_norm": 0.3727389872074127, "learning_rate": 8e-05, "loss": 1.8759, "step": 955 }, { "epoch": 0.10331784286177456, "grad_norm": 0.38414549827575684, "learning_rate": 8e-05, "loss": 1.8427, "step": 956 }, { "epoch": 0.10342591591916135, "grad_norm": 0.33417871594429016, "learning_rate": 8e-05, "loss": 1.7205, "step": 957 }, { "epoch": 0.10353398897654814, "grad_norm": 0.4005602300167084, "learning_rate": 8e-05, "loss": 1.9061, "step": 958 }, { "epoch": 0.10364206203393495, "grad_norm": 0.36298057436943054, "learning_rate": 8e-05, "loss": 1.5239, "step": 959 }, { "epoch": 0.10375013509132174, "grad_norm": 0.35376080870628357, "learning_rate": 8e-05, "loss": 1.7882, "step": 960 }, { "epoch": 0.10385820814870853, "grad_norm": 0.33489012718200684, "learning_rate": 8e-05, "loss": 1.6245, "step": 961 }, { "epoch": 0.10396628120609532, "grad_norm": 0.36470216512680054, "learning_rate": 8e-05, "loss": 1.6185, "step": 962 }, { "epoch": 0.10407435426348212, "grad_norm": 0.3187197744846344, "learning_rate": 8e-05, "loss": 1.6175, "step": 963 }, { "epoch": 0.10418242732086891, "grad_norm": 0.3826877176761627, "learning_rate": 8e-05, "loss": 1.9342, "step": 964 }, { "epoch": 0.1042905003782557, "grad_norm": 0.3329418897628784, "learning_rate": 8e-05, "loss": 1.6955, "step": 965 }, { "epoch": 0.10439857343564249, "grad_norm": 0.35413625836372375, "learning_rate": 8e-05, "loss": 1.9804, "step": 966 }, { "epoch": 0.10450664649302929, "grad_norm": 0.3780409097671509, "learning_rate": 8e-05, "loss": 1.793, "step": 967 }, { "epoch": 0.10461471955041608, "grad_norm": 0.34960803389549255, "learning_rate": 8e-05, "loss": 1.749, "step": 968 }, { "epoch": 0.10472279260780287, "grad_norm": 0.36611899733543396, "learning_rate": 8e-05, "loss": 1.8597, "step": 969 }, { "epoch": 0.10483086566518966, "grad_norm": 0.36785924434661865, "learning_rate": 8e-05, "loss": 1.7564, "step": 970 }, { "epoch": 0.10493893872257647, "grad_norm": 0.34815487265586853, "learning_rate": 8e-05, "loss": 1.6009, "step": 971 }, { "epoch": 0.10504701177996326, "grad_norm": 0.3491874933242798, "learning_rate": 8e-05, "loss": 1.8374, "step": 972 }, { "epoch": 0.10515508483735005, "grad_norm": 0.3539784550666809, "learning_rate": 8e-05, "loss": 1.7529, "step": 973 }, { "epoch": 0.10526315789473684, "grad_norm": 0.36112168431282043, "learning_rate": 8e-05, "loss": 1.8739, "step": 974 }, { "epoch": 0.10537123095212364, "grad_norm": 0.35476866364479065, "learning_rate": 8e-05, "loss": 1.7412, "step": 975 }, { "epoch": 0.10547930400951043, "grad_norm": 0.3740367293357849, "learning_rate": 8e-05, "loss": 1.8731, "step": 976 }, { "epoch": 0.10558737706689722, "grad_norm": 0.380718469619751, "learning_rate": 8e-05, "loss": 1.8771, "step": 977 }, { "epoch": 0.10569545012428401, "grad_norm": 0.3447098731994629, "learning_rate": 8e-05, "loss": 1.7702, "step": 978 }, { "epoch": 0.10580352318167081, "grad_norm": 0.3620442748069763, "learning_rate": 8e-05, "loss": 1.7765, "step": 979 }, { "epoch": 0.1059115962390576, "grad_norm": 0.3551279604434967, "learning_rate": 8e-05, "loss": 1.7793, "step": 980 }, { "epoch": 0.1060196692964444, "grad_norm": 0.3821079432964325, "learning_rate": 8e-05, "loss": 1.8377, "step": 981 }, { "epoch": 0.10612774235383118, "grad_norm": 0.3099254369735718, "learning_rate": 8e-05, "loss": 1.6299, "step": 982 }, { "epoch": 0.10623581541121799, "grad_norm": 0.3426049053668976, "learning_rate": 8e-05, "loss": 1.679, "step": 983 }, { "epoch": 0.10634388846860478, "grad_norm": 0.4244281053543091, "learning_rate": 8e-05, "loss": 1.9273, "step": 984 }, { "epoch": 0.10645196152599157, "grad_norm": 0.37145841121673584, "learning_rate": 8e-05, "loss": 1.8507, "step": 985 }, { "epoch": 0.10656003458337836, "grad_norm": 0.34629592299461365, "learning_rate": 8e-05, "loss": 1.7356, "step": 986 }, { "epoch": 0.10666810764076516, "grad_norm": 0.35548093914985657, "learning_rate": 8e-05, "loss": 1.6655, "step": 987 }, { "epoch": 0.10677618069815195, "grad_norm": 0.3681755065917969, "learning_rate": 8e-05, "loss": 1.7436, "step": 988 }, { "epoch": 0.10688425375553874, "grad_norm": 0.345228374004364, "learning_rate": 8e-05, "loss": 1.6541, "step": 989 }, { "epoch": 0.10699232681292553, "grad_norm": 0.36760804057121277, "learning_rate": 8e-05, "loss": 1.722, "step": 990 }, { "epoch": 0.10710039987031233, "grad_norm": 0.35668179392814636, "learning_rate": 8e-05, "loss": 1.7939, "step": 991 }, { "epoch": 0.10720847292769912, "grad_norm": 0.40337005257606506, "learning_rate": 8e-05, "loss": 2.0045, "step": 992 }, { "epoch": 0.10731654598508591, "grad_norm": 0.3598701059818268, "learning_rate": 8e-05, "loss": 1.6842, "step": 993 }, { "epoch": 0.1074246190424727, "grad_norm": 0.3536721467971802, "learning_rate": 8e-05, "loss": 1.7006, "step": 994 }, { "epoch": 0.10753269209985951, "grad_norm": 0.3558799624443054, "learning_rate": 8e-05, "loss": 1.7123, "step": 995 }, { "epoch": 0.1076407651572463, "grad_norm": 0.3420169949531555, "learning_rate": 8e-05, "loss": 1.7032, "step": 996 }, { "epoch": 0.10774883821463309, "grad_norm": 0.35889682173728943, "learning_rate": 8e-05, "loss": 1.906, "step": 997 }, { "epoch": 0.10785691127201988, "grad_norm": 0.3473713994026184, "learning_rate": 8e-05, "loss": 1.7758, "step": 998 }, { "epoch": 0.10796498432940668, "grad_norm": 0.3569095730781555, "learning_rate": 8e-05, "loss": 1.6352, "step": 999 }, { "epoch": 0.10807305738679347, "grad_norm": 0.36021849513053894, "learning_rate": 8e-05, "loss": 1.8736, "step": 1000 }, { "epoch": 0.10818113044418026, "grad_norm": 0.38916531205177307, "learning_rate": 8e-05, "loss": 1.9118, "step": 1001 }, { "epoch": 0.10828920350156707, "grad_norm": 0.3654513359069824, "learning_rate": 8e-05, "loss": 1.9466, "step": 1002 }, { "epoch": 0.10839727655895386, "grad_norm": 0.3699779212474823, "learning_rate": 8e-05, "loss": 1.7439, "step": 1003 }, { "epoch": 0.10850534961634065, "grad_norm": 0.37093672156333923, "learning_rate": 8e-05, "loss": 1.8799, "step": 1004 }, { "epoch": 0.10861342267372744, "grad_norm": 0.37014082074165344, "learning_rate": 8e-05, "loss": 1.8168, "step": 1005 }, { "epoch": 0.10872149573111424, "grad_norm": 0.36623871326446533, "learning_rate": 8e-05, "loss": 1.835, "step": 1006 }, { "epoch": 0.10882956878850103, "grad_norm": 0.34573739767074585, "learning_rate": 8e-05, "loss": 1.8492, "step": 1007 }, { "epoch": 0.10893764184588782, "grad_norm": 0.3408907949924469, "learning_rate": 8e-05, "loss": 1.6964, "step": 1008 }, { "epoch": 0.10904571490327461, "grad_norm": 0.34560516476631165, "learning_rate": 8e-05, "loss": 1.7529, "step": 1009 }, { "epoch": 0.10915378796066141, "grad_norm": 0.3811988830566406, "learning_rate": 8e-05, "loss": 1.9763, "step": 1010 }, { "epoch": 0.1092618610180482, "grad_norm": 0.4765642285346985, "learning_rate": 8e-05, "loss": 2.1039, "step": 1011 }, { "epoch": 0.109369934075435, "grad_norm": 0.34004926681518555, "learning_rate": 8e-05, "loss": 1.7518, "step": 1012 }, { "epoch": 0.10947800713282178, "grad_norm": 0.3730889558792114, "learning_rate": 8e-05, "loss": 1.897, "step": 1013 }, { "epoch": 0.10958608019020859, "grad_norm": 0.3418436050415039, "learning_rate": 8e-05, "loss": 1.7766, "step": 1014 }, { "epoch": 0.10969415324759538, "grad_norm": 0.324800580739975, "learning_rate": 8e-05, "loss": 1.7566, "step": 1015 }, { "epoch": 0.10980222630498217, "grad_norm": 0.3472938537597656, "learning_rate": 8e-05, "loss": 1.6911, "step": 1016 }, { "epoch": 0.10991029936236896, "grad_norm": 0.3487228453159332, "learning_rate": 8e-05, "loss": 1.6507, "step": 1017 }, { "epoch": 0.11001837241975576, "grad_norm": 0.3387090563774109, "learning_rate": 8e-05, "loss": 1.7957, "step": 1018 }, { "epoch": 0.11012644547714255, "grad_norm": 0.33071210980415344, "learning_rate": 8e-05, "loss": 1.7122, "step": 1019 }, { "epoch": 0.11023451853452934, "grad_norm": 0.35964852571487427, "learning_rate": 8e-05, "loss": 1.7012, "step": 1020 }, { "epoch": 0.11034259159191613, "grad_norm": 0.3740130364894867, "learning_rate": 8e-05, "loss": 1.7388, "step": 1021 }, { "epoch": 0.11045066464930293, "grad_norm": 0.34640559554100037, "learning_rate": 8e-05, "loss": 1.754, "step": 1022 }, { "epoch": 0.11055873770668972, "grad_norm": 0.33933576941490173, "learning_rate": 8e-05, "loss": 1.8113, "step": 1023 }, { "epoch": 0.11066681076407651, "grad_norm": 0.36161407828330994, "learning_rate": 8e-05, "loss": 1.7882, "step": 1024 }, { "epoch": 0.1107748838214633, "grad_norm": 0.3854907155036926, "learning_rate": 8e-05, "loss": 1.8705, "step": 1025 }, { "epoch": 0.11088295687885011, "grad_norm": 0.35772159695625305, "learning_rate": 8e-05, "loss": 1.6783, "step": 1026 }, { "epoch": 0.1109910299362369, "grad_norm": 0.3468589782714844, "learning_rate": 8e-05, "loss": 1.7651, "step": 1027 }, { "epoch": 0.11109910299362369, "grad_norm": 0.3650378882884979, "learning_rate": 8e-05, "loss": 1.7646, "step": 1028 }, { "epoch": 0.11120717605101048, "grad_norm": 0.3954007923603058, "learning_rate": 8e-05, "loss": 1.6899, "step": 1029 }, { "epoch": 0.11131524910839728, "grad_norm": 0.34445634484291077, "learning_rate": 8e-05, "loss": 1.7073, "step": 1030 }, { "epoch": 0.11142332216578407, "grad_norm": 0.3457842469215393, "learning_rate": 8e-05, "loss": 1.8061, "step": 1031 }, { "epoch": 0.11153139522317086, "grad_norm": 0.3583963215351105, "learning_rate": 8e-05, "loss": 1.8769, "step": 1032 }, { "epoch": 0.11163946828055765, "grad_norm": 0.3492564558982849, "learning_rate": 8e-05, "loss": 1.7665, "step": 1033 }, { "epoch": 0.11174754133794446, "grad_norm": 0.3324894607067108, "learning_rate": 8e-05, "loss": 1.7425, "step": 1034 }, { "epoch": 0.11185561439533125, "grad_norm": 0.34477636218070984, "learning_rate": 8e-05, "loss": 1.7223, "step": 1035 }, { "epoch": 0.11196368745271804, "grad_norm": 0.35887598991394043, "learning_rate": 8e-05, "loss": 1.8317, "step": 1036 }, { "epoch": 0.11207176051010483, "grad_norm": 0.359162300825119, "learning_rate": 8e-05, "loss": 1.8598, "step": 1037 }, { "epoch": 0.11217983356749163, "grad_norm": 0.37809160351753235, "learning_rate": 8e-05, "loss": 1.8609, "step": 1038 }, { "epoch": 0.11228790662487842, "grad_norm": 0.34972062706947327, "learning_rate": 8e-05, "loss": 1.6576, "step": 1039 }, { "epoch": 0.11239597968226521, "grad_norm": 0.3460668623447418, "learning_rate": 8e-05, "loss": 1.6272, "step": 1040 }, { "epoch": 0.112504052739652, "grad_norm": 0.3861105144023895, "learning_rate": 8e-05, "loss": 2.0128, "step": 1041 }, { "epoch": 0.1126121257970388, "grad_norm": 0.35771337151527405, "learning_rate": 8e-05, "loss": 1.8703, "step": 1042 }, { "epoch": 0.1127201988544256, "grad_norm": 0.335125207901001, "learning_rate": 8e-05, "loss": 1.8183, "step": 1043 }, { "epoch": 0.11282827191181238, "grad_norm": 0.3440723121166229, "learning_rate": 8e-05, "loss": 1.8641, "step": 1044 }, { "epoch": 0.11293634496919917, "grad_norm": 0.37276747822761536, "learning_rate": 8e-05, "loss": 1.773, "step": 1045 }, { "epoch": 0.11304441802658598, "grad_norm": 0.34359925985336304, "learning_rate": 8e-05, "loss": 1.7484, "step": 1046 }, { "epoch": 0.11315249108397277, "grad_norm": 0.33626818656921387, "learning_rate": 8e-05, "loss": 1.7244, "step": 1047 }, { "epoch": 0.11326056414135956, "grad_norm": 0.36569783091545105, "learning_rate": 8e-05, "loss": 1.8492, "step": 1048 }, { "epoch": 0.11336863719874635, "grad_norm": 0.33529019355773926, "learning_rate": 8e-05, "loss": 1.6912, "step": 1049 }, { "epoch": 0.11347671025613315, "grad_norm": 0.33651697635650635, "learning_rate": 8e-05, "loss": 1.5635, "step": 1050 }, { "epoch": 0.11358478331351994, "grad_norm": 0.3362615704536438, "learning_rate": 8e-05, "loss": 1.6028, "step": 1051 }, { "epoch": 0.11369285637090673, "grad_norm": 0.33662670850753784, "learning_rate": 8e-05, "loss": 1.733, "step": 1052 }, { "epoch": 0.11380092942829352, "grad_norm": 0.3547325134277344, "learning_rate": 8e-05, "loss": 1.73, "step": 1053 }, { "epoch": 0.11390900248568032, "grad_norm": 0.33198556303977966, "learning_rate": 8e-05, "loss": 1.6468, "step": 1054 }, { "epoch": 0.11401707554306711, "grad_norm": 0.33353909850120544, "learning_rate": 8e-05, "loss": 1.7448, "step": 1055 }, { "epoch": 0.1141251486004539, "grad_norm": 0.3565393388271332, "learning_rate": 8e-05, "loss": 1.8502, "step": 1056 }, { "epoch": 0.1142332216578407, "grad_norm": 0.37357306480407715, "learning_rate": 8e-05, "loss": 1.7812, "step": 1057 }, { "epoch": 0.1143412947152275, "grad_norm": 0.37155699729919434, "learning_rate": 8e-05, "loss": 1.8798, "step": 1058 }, { "epoch": 0.11444936777261429, "grad_norm": 0.3555389642715454, "learning_rate": 8e-05, "loss": 1.8011, "step": 1059 }, { "epoch": 0.11455744083000108, "grad_norm": 0.3473822772502899, "learning_rate": 8e-05, "loss": 1.7081, "step": 1060 }, { "epoch": 0.11466551388738787, "grad_norm": 0.34955134987831116, "learning_rate": 8e-05, "loss": 1.7404, "step": 1061 }, { "epoch": 0.11477358694477467, "grad_norm": 0.36052483320236206, "learning_rate": 8e-05, "loss": 1.7622, "step": 1062 }, { "epoch": 0.11488166000216146, "grad_norm": 0.32584789395332336, "learning_rate": 8e-05, "loss": 1.6502, "step": 1063 }, { "epoch": 0.11498973305954825, "grad_norm": 0.3796273171901703, "learning_rate": 8e-05, "loss": 1.7895, "step": 1064 }, { "epoch": 0.11509780611693504, "grad_norm": 0.3480946123600006, "learning_rate": 8e-05, "loss": 1.8156, "step": 1065 }, { "epoch": 0.11520587917432185, "grad_norm": 0.3912273347377777, "learning_rate": 8e-05, "loss": 1.9485, "step": 1066 }, { "epoch": 0.11531395223170864, "grad_norm": 0.3587355613708496, "learning_rate": 8e-05, "loss": 1.469, "step": 1067 }, { "epoch": 0.11542202528909543, "grad_norm": 0.36560285091400146, "learning_rate": 8e-05, "loss": 1.9126, "step": 1068 }, { "epoch": 0.11553009834648222, "grad_norm": 0.34391555190086365, "learning_rate": 8e-05, "loss": 1.687, "step": 1069 }, { "epoch": 0.11563817140386902, "grad_norm": 0.3561718165874481, "learning_rate": 8e-05, "loss": 1.7612, "step": 1070 }, { "epoch": 0.11574624446125581, "grad_norm": 0.39095380902290344, "learning_rate": 8e-05, "loss": 1.9475, "step": 1071 }, { "epoch": 0.1158543175186426, "grad_norm": 0.3762200176715851, "learning_rate": 8e-05, "loss": 1.906, "step": 1072 }, { "epoch": 0.11596239057602939, "grad_norm": 0.3578304052352905, "learning_rate": 8e-05, "loss": 1.7851, "step": 1073 }, { "epoch": 0.1160704636334162, "grad_norm": 0.3298514485359192, "learning_rate": 8e-05, "loss": 1.6766, "step": 1074 }, { "epoch": 0.11617853669080298, "grad_norm": 0.3591746687889099, "learning_rate": 8e-05, "loss": 1.8481, "step": 1075 }, { "epoch": 0.11628660974818977, "grad_norm": 0.3621608018875122, "learning_rate": 8e-05, "loss": 1.7246, "step": 1076 }, { "epoch": 0.11639468280557656, "grad_norm": 0.3578890562057495, "learning_rate": 8e-05, "loss": 1.7048, "step": 1077 }, { "epoch": 0.11650275586296337, "grad_norm": 0.4235458970069885, "learning_rate": 8e-05, "loss": 1.7364, "step": 1078 }, { "epoch": 0.11661082892035016, "grad_norm": 0.35560694336891174, "learning_rate": 8e-05, "loss": 1.9208, "step": 1079 }, { "epoch": 0.11671890197773695, "grad_norm": 0.3598167896270752, "learning_rate": 8e-05, "loss": 1.8029, "step": 1080 }, { "epoch": 0.11682697503512374, "grad_norm": 0.37738943099975586, "learning_rate": 8e-05, "loss": 1.9841, "step": 1081 }, { "epoch": 0.11693504809251054, "grad_norm": 0.3868809938430786, "learning_rate": 8e-05, "loss": 1.9314, "step": 1082 }, { "epoch": 0.11704312114989733, "grad_norm": 0.3519797623157501, "learning_rate": 8e-05, "loss": 1.7351, "step": 1083 }, { "epoch": 0.11715119420728412, "grad_norm": 0.33653125166893005, "learning_rate": 8e-05, "loss": 1.6673, "step": 1084 }, { "epoch": 0.11725926726467092, "grad_norm": 0.3813892602920532, "learning_rate": 8e-05, "loss": 1.9151, "step": 1085 }, { "epoch": 0.11736734032205771, "grad_norm": 0.36971303820610046, "learning_rate": 8e-05, "loss": 1.772, "step": 1086 }, { "epoch": 0.1174754133794445, "grad_norm": 0.35533636808395386, "learning_rate": 8e-05, "loss": 1.7835, "step": 1087 }, { "epoch": 0.1175834864368313, "grad_norm": 0.36718428134918213, "learning_rate": 8e-05, "loss": 1.7383, "step": 1088 }, { "epoch": 0.1176915594942181, "grad_norm": 0.3542182743549347, "learning_rate": 8e-05, "loss": 1.7868, "step": 1089 }, { "epoch": 0.11779963255160489, "grad_norm": 0.42764219641685486, "learning_rate": 8e-05, "loss": 1.6949, "step": 1090 }, { "epoch": 0.11790770560899168, "grad_norm": 0.33852437138557434, "learning_rate": 8e-05, "loss": 1.7044, "step": 1091 }, { "epoch": 0.11801577866637847, "grad_norm": 0.37063103914260864, "learning_rate": 8e-05, "loss": 1.8017, "step": 1092 }, { "epoch": 0.11812385172376527, "grad_norm": 0.3522622287273407, "learning_rate": 8e-05, "loss": 1.7746, "step": 1093 }, { "epoch": 0.11823192478115206, "grad_norm": 0.36144259572029114, "learning_rate": 8e-05, "loss": 1.859, "step": 1094 }, { "epoch": 0.11833999783853885, "grad_norm": 0.36767423152923584, "learning_rate": 8e-05, "loss": 1.8105, "step": 1095 }, { "epoch": 0.11844807089592564, "grad_norm": 0.36035487055778503, "learning_rate": 8e-05, "loss": 1.7141, "step": 1096 }, { "epoch": 0.11855614395331245, "grad_norm": 0.38611260056495667, "learning_rate": 8e-05, "loss": 1.7955, "step": 1097 }, { "epoch": 0.11866421701069924, "grad_norm": 0.336543470621109, "learning_rate": 8e-05, "loss": 1.8048, "step": 1098 }, { "epoch": 0.11877229006808603, "grad_norm": 0.34846892952919006, "learning_rate": 8e-05, "loss": 1.7128, "step": 1099 }, { "epoch": 0.11888036312547282, "grad_norm": 0.37330928444862366, "learning_rate": 8e-05, "loss": 1.7228, "step": 1100 }, { "epoch": 0.11898843618285962, "grad_norm": 0.3681880533695221, "learning_rate": 8e-05, "loss": 1.7345, "step": 1101 }, { "epoch": 0.11909650924024641, "grad_norm": 0.3820163905620575, "learning_rate": 8e-05, "loss": 1.9009, "step": 1102 }, { "epoch": 0.1192045822976332, "grad_norm": 0.3882463276386261, "learning_rate": 8e-05, "loss": 1.8318, "step": 1103 }, { "epoch": 0.11931265535501999, "grad_norm": 0.398868203163147, "learning_rate": 8e-05, "loss": 1.8965, "step": 1104 }, { "epoch": 0.11942072841240679, "grad_norm": 0.3507888913154602, "learning_rate": 8e-05, "loss": 1.5984, "step": 1105 }, { "epoch": 0.11952880146979358, "grad_norm": 0.36435365676879883, "learning_rate": 8e-05, "loss": 1.8051, "step": 1106 }, { "epoch": 0.11963687452718037, "grad_norm": 0.3470079004764557, "learning_rate": 8e-05, "loss": 1.8356, "step": 1107 }, { "epoch": 0.11974494758456716, "grad_norm": 0.441213458776474, "learning_rate": 8e-05, "loss": 1.7817, "step": 1108 }, { "epoch": 0.11985302064195397, "grad_norm": 0.35217058658599854, "learning_rate": 8e-05, "loss": 1.7039, "step": 1109 }, { "epoch": 0.11996109369934076, "grad_norm": 0.3345927894115448, "learning_rate": 8e-05, "loss": 1.7784, "step": 1110 }, { "epoch": 0.12006916675672755, "grad_norm": 0.37604644894599915, "learning_rate": 8e-05, "loss": 1.7291, "step": 1111 }, { "epoch": 0.12017723981411434, "grad_norm": 0.40729671716690063, "learning_rate": 8e-05, "loss": 1.8353, "step": 1112 }, { "epoch": 0.12028531287150114, "grad_norm": 0.3934181332588196, "learning_rate": 8e-05, "loss": 1.8137, "step": 1113 }, { "epoch": 0.12039338592888793, "grad_norm": 0.406771719455719, "learning_rate": 8e-05, "loss": 1.9328, "step": 1114 }, { "epoch": 0.12050145898627472, "grad_norm": 0.37399056553840637, "learning_rate": 8e-05, "loss": 1.767, "step": 1115 }, { "epoch": 0.12060953204366151, "grad_norm": 0.38063445687294006, "learning_rate": 8e-05, "loss": 1.752, "step": 1116 }, { "epoch": 0.12071760510104831, "grad_norm": 0.4235222339630127, "learning_rate": 8e-05, "loss": 1.749, "step": 1117 }, { "epoch": 0.1208256781584351, "grad_norm": 0.3890572786331177, "learning_rate": 8e-05, "loss": 1.9588, "step": 1118 }, { "epoch": 0.1209337512158219, "grad_norm": 0.3706381916999817, "learning_rate": 8e-05, "loss": 1.7294, "step": 1119 }, { "epoch": 0.12104182427320868, "grad_norm": 0.374380499124527, "learning_rate": 8e-05, "loss": 1.7579, "step": 1120 }, { "epoch": 0.12114989733059549, "grad_norm": 0.361548513174057, "learning_rate": 8e-05, "loss": 1.6716, "step": 1121 }, { "epoch": 0.12125797038798228, "grad_norm": 0.36958298087120056, "learning_rate": 8e-05, "loss": 1.739, "step": 1122 }, { "epoch": 0.12136604344536907, "grad_norm": 0.3657321333885193, "learning_rate": 8e-05, "loss": 1.8152, "step": 1123 }, { "epoch": 0.12147411650275586, "grad_norm": 0.35615840554237366, "learning_rate": 8e-05, "loss": 1.6348, "step": 1124 }, { "epoch": 0.12158218956014266, "grad_norm": 0.3991694748401642, "learning_rate": 8e-05, "loss": 1.967, "step": 1125 }, { "epoch": 0.12169026261752945, "grad_norm": 0.3788558542728424, "learning_rate": 8e-05, "loss": 1.8675, "step": 1126 }, { "epoch": 0.12179833567491624, "grad_norm": 0.3795939087867737, "learning_rate": 8e-05, "loss": 1.9319, "step": 1127 }, { "epoch": 0.12190640873230303, "grad_norm": 0.36552003026008606, "learning_rate": 8e-05, "loss": 1.773, "step": 1128 }, { "epoch": 0.12201448178968984, "grad_norm": 0.3401627838611603, "learning_rate": 8e-05, "loss": 1.6467, "step": 1129 }, { "epoch": 0.12212255484707663, "grad_norm": 0.3407789170742035, "learning_rate": 8e-05, "loss": 1.5765, "step": 1130 }, { "epoch": 0.12223062790446342, "grad_norm": 0.35426127910614014, "learning_rate": 8e-05, "loss": 1.6499, "step": 1131 }, { "epoch": 0.1223387009618502, "grad_norm": 0.38160404562950134, "learning_rate": 8e-05, "loss": 1.9214, "step": 1132 }, { "epoch": 0.12244677401923701, "grad_norm": 0.35989806056022644, "learning_rate": 8e-05, "loss": 1.8379, "step": 1133 }, { "epoch": 0.1225548470766238, "grad_norm": 0.41352027654647827, "learning_rate": 8e-05, "loss": 1.7321, "step": 1134 }, { "epoch": 0.12266292013401059, "grad_norm": 0.36961716413497925, "learning_rate": 8e-05, "loss": 1.9075, "step": 1135 }, { "epoch": 0.12277099319139738, "grad_norm": 0.3447885513305664, "learning_rate": 8e-05, "loss": 1.8646, "step": 1136 }, { "epoch": 0.12287906624878418, "grad_norm": 0.3339730203151703, "learning_rate": 8e-05, "loss": 1.7248, "step": 1137 }, { "epoch": 0.12298713930617097, "grad_norm": 0.33837664127349854, "learning_rate": 8e-05, "loss": 1.6525, "step": 1138 }, { "epoch": 0.12309521236355776, "grad_norm": 0.40510615706443787, "learning_rate": 8e-05, "loss": 1.9152, "step": 1139 }, { "epoch": 0.12320328542094455, "grad_norm": 0.3553943932056427, "learning_rate": 8e-05, "loss": 1.8305, "step": 1140 }, { "epoch": 0.12331135847833136, "grad_norm": 0.3435276746749878, "learning_rate": 8e-05, "loss": 1.7215, "step": 1141 }, { "epoch": 0.12341943153571815, "grad_norm": 0.3574138879776001, "learning_rate": 8e-05, "loss": 1.7144, "step": 1142 }, { "epoch": 0.12352750459310494, "grad_norm": 0.3462582528591156, "learning_rate": 8e-05, "loss": 1.7756, "step": 1143 }, { "epoch": 0.12363557765049173, "grad_norm": 0.3465636372566223, "learning_rate": 8e-05, "loss": 1.7971, "step": 1144 }, { "epoch": 0.12374365070787853, "grad_norm": 0.34524717926979065, "learning_rate": 8e-05, "loss": 1.7362, "step": 1145 }, { "epoch": 0.12385172376526532, "grad_norm": 0.3610204756259918, "learning_rate": 8e-05, "loss": 1.8568, "step": 1146 }, { "epoch": 0.12395979682265211, "grad_norm": 0.3694356083869934, "learning_rate": 8e-05, "loss": 1.8118, "step": 1147 }, { "epoch": 0.1240678698800389, "grad_norm": 0.3656586706638336, "learning_rate": 8e-05, "loss": 1.963, "step": 1148 }, { "epoch": 0.1241759429374257, "grad_norm": 0.40215447545051575, "learning_rate": 8e-05, "loss": 1.9987, "step": 1149 }, { "epoch": 0.1242840159948125, "grad_norm": 0.3576542139053345, "learning_rate": 8e-05, "loss": 1.867, "step": 1150 }, { "epoch": 0.12439208905219928, "grad_norm": 0.3616178333759308, "learning_rate": 8e-05, "loss": 1.7104, "step": 1151 }, { "epoch": 0.12450016210958607, "grad_norm": 0.3752042353153229, "learning_rate": 8e-05, "loss": 2.0495, "step": 1152 }, { "epoch": 0.12460823516697288, "grad_norm": 0.3590239882469177, "learning_rate": 8e-05, "loss": 1.7342, "step": 1153 }, { "epoch": 0.12471630822435967, "grad_norm": 0.3245165944099426, "learning_rate": 8e-05, "loss": 1.547, "step": 1154 }, { "epoch": 0.12482438128174646, "grad_norm": 0.37018802762031555, "learning_rate": 8e-05, "loss": 1.7599, "step": 1155 }, { "epoch": 0.12493245433913325, "grad_norm": 0.33771440386772156, "learning_rate": 8e-05, "loss": 1.8027, "step": 1156 }, { "epoch": 0.12504052739652005, "grad_norm": 0.34656190872192383, "learning_rate": 8e-05, "loss": 1.9092, "step": 1157 }, { "epoch": 0.12514860045390683, "grad_norm": 0.3510362505912781, "learning_rate": 8e-05, "loss": 1.9008, "step": 1158 }, { "epoch": 0.12525667351129363, "grad_norm": 0.3645678758621216, "learning_rate": 8e-05, "loss": 1.8581, "step": 1159 }, { "epoch": 0.12536474656868044, "grad_norm": 0.32818347215652466, "learning_rate": 8e-05, "loss": 1.6762, "step": 1160 }, { "epoch": 0.1254728196260672, "grad_norm": 0.3408606946468353, "learning_rate": 8e-05, "loss": 1.7287, "step": 1161 }, { "epoch": 0.12558089268345402, "grad_norm": 0.3740525245666504, "learning_rate": 8e-05, "loss": 1.7846, "step": 1162 }, { "epoch": 0.12568896574084082, "grad_norm": 0.35428285598754883, "learning_rate": 8e-05, "loss": 1.7405, "step": 1163 }, { "epoch": 0.1257970387982276, "grad_norm": 0.36056891083717346, "learning_rate": 8e-05, "loss": 1.7253, "step": 1164 }, { "epoch": 0.1259051118556144, "grad_norm": 0.3450113832950592, "learning_rate": 8e-05, "loss": 1.5222, "step": 1165 }, { "epoch": 0.12601318491300117, "grad_norm": 0.3771056532859802, "learning_rate": 8e-05, "loss": 1.8111, "step": 1166 }, { "epoch": 0.12612125797038798, "grad_norm": 0.34107789397239685, "learning_rate": 8e-05, "loss": 1.7258, "step": 1167 }, { "epoch": 0.12622933102777478, "grad_norm": 0.3527986705303192, "learning_rate": 8e-05, "loss": 1.7354, "step": 1168 }, { "epoch": 0.12633740408516156, "grad_norm": 0.3723733127117157, "learning_rate": 8e-05, "loss": 1.906, "step": 1169 }, { "epoch": 0.12644547714254836, "grad_norm": 0.36626508831977844, "learning_rate": 8e-05, "loss": 1.7188, "step": 1170 }, { "epoch": 0.12655355019993517, "grad_norm": 0.3760107755661011, "learning_rate": 8e-05, "loss": 1.9295, "step": 1171 }, { "epoch": 0.12666162325732194, "grad_norm": 0.32605016231536865, "learning_rate": 8e-05, "loss": 1.726, "step": 1172 }, { "epoch": 0.12676969631470875, "grad_norm": 0.389241099357605, "learning_rate": 8e-05, "loss": 1.6831, "step": 1173 }, { "epoch": 0.12687776937209555, "grad_norm": 0.3753553330898285, "learning_rate": 8e-05, "loss": 1.7441, "step": 1174 }, { "epoch": 0.12698584242948233, "grad_norm": 0.36482882499694824, "learning_rate": 8e-05, "loss": 1.7286, "step": 1175 }, { "epoch": 0.12709391548686913, "grad_norm": 0.3554099500179291, "learning_rate": 8e-05, "loss": 1.7884, "step": 1176 }, { "epoch": 0.1272019885442559, "grad_norm": 0.33709585666656494, "learning_rate": 8e-05, "loss": 1.6052, "step": 1177 }, { "epoch": 0.1273100616016427, "grad_norm": 0.3496922552585602, "learning_rate": 8e-05, "loss": 1.586, "step": 1178 }, { "epoch": 0.1274181346590295, "grad_norm": 0.3904522657394409, "learning_rate": 8e-05, "loss": 1.9564, "step": 1179 }, { "epoch": 0.1275262077164163, "grad_norm": 0.36048629879951477, "learning_rate": 8e-05, "loss": 1.9061, "step": 1180 }, { "epoch": 0.1276342807738031, "grad_norm": 0.3584705889225006, "learning_rate": 8e-05, "loss": 1.7315, "step": 1181 }, { "epoch": 0.1277423538311899, "grad_norm": 0.34056371450424194, "learning_rate": 8e-05, "loss": 1.6468, "step": 1182 }, { "epoch": 0.12785042688857667, "grad_norm": 0.37036845088005066, "learning_rate": 8e-05, "loss": 1.7558, "step": 1183 }, { "epoch": 0.12795849994596348, "grad_norm": 0.3162347972393036, "learning_rate": 8e-05, "loss": 1.5075, "step": 1184 }, { "epoch": 0.12806657300335025, "grad_norm": 0.3467709422111511, "learning_rate": 8e-05, "loss": 1.7704, "step": 1185 }, { "epoch": 0.12817464606073706, "grad_norm": 0.3360818028450012, "learning_rate": 8e-05, "loss": 1.7762, "step": 1186 }, { "epoch": 0.12828271911812386, "grad_norm": 0.35458385944366455, "learning_rate": 8e-05, "loss": 1.7368, "step": 1187 }, { "epoch": 0.12839079217551064, "grad_norm": 0.3742386996746063, "learning_rate": 8e-05, "loss": 1.7366, "step": 1188 }, { "epoch": 0.12849886523289744, "grad_norm": 0.3503303825855255, "learning_rate": 8e-05, "loss": 1.7712, "step": 1189 }, { "epoch": 0.12860693829028425, "grad_norm": 0.3646455407142639, "learning_rate": 8e-05, "loss": 1.7931, "step": 1190 }, { "epoch": 0.12871501134767102, "grad_norm": 0.37325340509414673, "learning_rate": 8e-05, "loss": 1.9369, "step": 1191 }, { "epoch": 0.12882308440505782, "grad_norm": 0.3590868413448334, "learning_rate": 8e-05, "loss": 1.7607, "step": 1192 }, { "epoch": 0.1289311574624446, "grad_norm": 0.3444075286388397, "learning_rate": 8e-05, "loss": 1.7446, "step": 1193 }, { "epoch": 0.1290392305198314, "grad_norm": 0.34096190333366394, "learning_rate": 8e-05, "loss": 1.8142, "step": 1194 }, { "epoch": 0.1291473035772182, "grad_norm": 0.35368281602859497, "learning_rate": 8e-05, "loss": 1.8032, "step": 1195 }, { "epoch": 0.12925537663460498, "grad_norm": 0.37291258573532104, "learning_rate": 8e-05, "loss": 1.8411, "step": 1196 }, { "epoch": 0.1293634496919918, "grad_norm": 0.3850989043712616, "learning_rate": 8e-05, "loss": 1.8738, "step": 1197 }, { "epoch": 0.1294715227493786, "grad_norm": 0.32650718092918396, "learning_rate": 8e-05, "loss": 1.6912, "step": 1198 }, { "epoch": 0.12957959580676537, "grad_norm": 0.40874555706977844, "learning_rate": 8e-05, "loss": 1.9786, "step": 1199 }, { "epoch": 0.12968766886415217, "grad_norm": 0.3875872492790222, "learning_rate": 8e-05, "loss": 1.837, "step": 1200 }, { "epoch": 0.12979574192153895, "grad_norm": 0.37259984016418457, "learning_rate": 8e-05, "loss": 1.9103, "step": 1201 }, { "epoch": 0.12990381497892575, "grad_norm": 0.36838170886039734, "learning_rate": 8e-05, "loss": 1.8204, "step": 1202 }, { "epoch": 0.13001188803631256, "grad_norm": 0.37012097239494324, "learning_rate": 8e-05, "loss": 1.8511, "step": 1203 }, { "epoch": 0.13011996109369933, "grad_norm": 0.3868945837020874, "learning_rate": 8e-05, "loss": 1.875, "step": 1204 }, { "epoch": 0.13022803415108614, "grad_norm": 0.3822103440761566, "learning_rate": 8e-05, "loss": 1.6806, "step": 1205 }, { "epoch": 0.13033610720847294, "grad_norm": 0.39760133624076843, "learning_rate": 8e-05, "loss": 1.8603, "step": 1206 }, { "epoch": 0.13044418026585972, "grad_norm": 0.3762820065021515, "learning_rate": 8e-05, "loss": 1.6615, "step": 1207 }, { "epoch": 0.13055225332324652, "grad_norm": 0.3529280424118042, "learning_rate": 8e-05, "loss": 1.8068, "step": 1208 }, { "epoch": 0.1306603263806333, "grad_norm": 0.3459187150001526, "learning_rate": 8e-05, "loss": 1.6836, "step": 1209 }, { "epoch": 0.1307683994380201, "grad_norm": 0.37171846628189087, "learning_rate": 8e-05, "loss": 1.7559, "step": 1210 }, { "epoch": 0.1308764724954069, "grad_norm": 0.32489991188049316, "learning_rate": 8e-05, "loss": 1.7134, "step": 1211 }, { "epoch": 0.13098454555279368, "grad_norm": 0.341099351644516, "learning_rate": 8e-05, "loss": 1.6784, "step": 1212 }, { "epoch": 0.13109261861018048, "grad_norm": 0.35070812702178955, "learning_rate": 8e-05, "loss": 1.8147, "step": 1213 }, { "epoch": 0.1312006916675673, "grad_norm": 0.3758661150932312, "learning_rate": 8e-05, "loss": 1.8404, "step": 1214 }, { "epoch": 0.13130876472495406, "grad_norm": 0.3853369653224945, "learning_rate": 8e-05, "loss": 1.7135, "step": 1215 }, { "epoch": 0.13141683778234087, "grad_norm": 0.48795756697654724, "learning_rate": 8e-05, "loss": 2.0874, "step": 1216 }, { "epoch": 0.13152491083972764, "grad_norm": 0.3734152615070343, "learning_rate": 8e-05, "loss": 1.8706, "step": 1217 }, { "epoch": 0.13163298389711445, "grad_norm": 0.39112815260887146, "learning_rate": 8e-05, "loss": 1.7548, "step": 1218 }, { "epoch": 0.13174105695450125, "grad_norm": 0.3530450165271759, "learning_rate": 8e-05, "loss": 1.7709, "step": 1219 }, { "epoch": 0.13184913001188803, "grad_norm": 0.3932957947254181, "learning_rate": 8e-05, "loss": 2.0326, "step": 1220 }, { "epoch": 0.13195720306927483, "grad_norm": 0.34830108284950256, "learning_rate": 8e-05, "loss": 1.8511, "step": 1221 }, { "epoch": 0.13206527612666163, "grad_norm": 0.42291244864463806, "learning_rate": 8e-05, "loss": 1.9159, "step": 1222 }, { "epoch": 0.1321733491840484, "grad_norm": 0.35291528701782227, "learning_rate": 8e-05, "loss": 1.8812, "step": 1223 }, { "epoch": 0.13228142224143521, "grad_norm": 0.38225409388542175, "learning_rate": 8e-05, "loss": 1.767, "step": 1224 }, { "epoch": 0.132389495298822, "grad_norm": 0.3345004916191101, "learning_rate": 8e-05, "loss": 1.6906, "step": 1225 }, { "epoch": 0.1324975683562088, "grad_norm": 0.3450838327407837, "learning_rate": 8e-05, "loss": 1.6551, "step": 1226 }, { "epoch": 0.1326056414135956, "grad_norm": 0.3691742420196533, "learning_rate": 8e-05, "loss": 1.8745, "step": 1227 }, { "epoch": 0.13271371447098237, "grad_norm": 0.3605695366859436, "learning_rate": 8e-05, "loss": 1.8219, "step": 1228 }, { "epoch": 0.13282178752836918, "grad_norm": 0.3561975061893463, "learning_rate": 8e-05, "loss": 1.7542, "step": 1229 }, { "epoch": 0.13292986058575598, "grad_norm": 0.3594520688056946, "learning_rate": 8e-05, "loss": 1.5149, "step": 1230 }, { "epoch": 0.13303793364314276, "grad_norm": 0.3900435268878937, "learning_rate": 8e-05, "loss": 1.8228, "step": 1231 }, { "epoch": 0.13314600670052956, "grad_norm": 0.3477350175380707, "learning_rate": 8e-05, "loss": 1.7469, "step": 1232 }, { "epoch": 0.13325407975791634, "grad_norm": 0.36821088194847107, "learning_rate": 8e-05, "loss": 1.8097, "step": 1233 }, { "epoch": 0.13336215281530314, "grad_norm": 0.3814113140106201, "learning_rate": 8e-05, "loss": 1.9105, "step": 1234 }, { "epoch": 0.13347022587268995, "grad_norm": 0.3567351996898651, "learning_rate": 8e-05, "loss": 1.8162, "step": 1235 }, { "epoch": 0.13357829893007672, "grad_norm": 0.38173460960388184, "learning_rate": 8e-05, "loss": 1.9196, "step": 1236 }, { "epoch": 0.13368637198746353, "grad_norm": 0.3624939024448395, "learning_rate": 8e-05, "loss": 1.8586, "step": 1237 }, { "epoch": 0.13379444504485033, "grad_norm": 0.34047120809555054, "learning_rate": 8e-05, "loss": 1.6708, "step": 1238 }, { "epoch": 0.1339025181022371, "grad_norm": 0.3519713282585144, "learning_rate": 8e-05, "loss": 1.7616, "step": 1239 }, { "epoch": 0.1340105911596239, "grad_norm": 0.37031808495521545, "learning_rate": 8e-05, "loss": 1.9764, "step": 1240 }, { "epoch": 0.13411866421701069, "grad_norm": 0.3479854464530945, "learning_rate": 8e-05, "loss": 1.5646, "step": 1241 }, { "epoch": 0.1342267372743975, "grad_norm": 0.3301011025905609, "learning_rate": 8e-05, "loss": 1.6588, "step": 1242 }, { "epoch": 0.1343348103317843, "grad_norm": 0.359833300113678, "learning_rate": 8e-05, "loss": 1.8077, "step": 1243 }, { "epoch": 0.13444288338917107, "grad_norm": 0.3639060854911804, "learning_rate": 8e-05, "loss": 1.7188, "step": 1244 }, { "epoch": 0.13455095644655787, "grad_norm": 0.3266666531562805, "learning_rate": 8e-05, "loss": 1.6254, "step": 1245 }, { "epoch": 0.13465902950394468, "grad_norm": 0.37093308568000793, "learning_rate": 8e-05, "loss": 1.758, "step": 1246 }, { "epoch": 0.13476710256133145, "grad_norm": 0.3673769235610962, "learning_rate": 8e-05, "loss": 1.7933, "step": 1247 }, { "epoch": 0.13487517561871826, "grad_norm": 0.35311612486839294, "learning_rate": 8e-05, "loss": 1.8474, "step": 1248 }, { "epoch": 0.13498324867610503, "grad_norm": 0.3570164144039154, "learning_rate": 8e-05, "loss": 1.8897, "step": 1249 }, { "epoch": 0.13509132173349184, "grad_norm": 0.3754791021347046, "learning_rate": 8e-05, "loss": 1.6946, "step": 1250 }, { "epoch": 0.13519939479087864, "grad_norm": 0.3997972011566162, "learning_rate": 8e-05, "loss": 1.8581, "step": 1251 }, { "epoch": 0.13530746784826542, "grad_norm": 0.39735230803489685, "learning_rate": 8e-05, "loss": 2.0488, "step": 1252 }, { "epoch": 0.13541554090565222, "grad_norm": 0.3601343035697937, "learning_rate": 8e-05, "loss": 1.5946, "step": 1253 }, { "epoch": 0.13552361396303902, "grad_norm": 0.36464381217956543, "learning_rate": 8e-05, "loss": 1.786, "step": 1254 }, { "epoch": 0.1356316870204258, "grad_norm": 0.3686005175113678, "learning_rate": 8e-05, "loss": 1.8295, "step": 1255 }, { "epoch": 0.1357397600778126, "grad_norm": 0.41247066855430603, "learning_rate": 8e-05, "loss": 1.9405, "step": 1256 }, { "epoch": 0.1358478331351994, "grad_norm": 0.3546912968158722, "learning_rate": 8e-05, "loss": 1.6884, "step": 1257 }, { "epoch": 0.13595590619258618, "grad_norm": 0.396194189786911, "learning_rate": 8e-05, "loss": 1.8007, "step": 1258 }, { "epoch": 0.136063979249973, "grad_norm": 0.3414531350135803, "learning_rate": 8e-05, "loss": 1.7376, "step": 1259 }, { "epoch": 0.13617205230735976, "grad_norm": 0.365448534488678, "learning_rate": 8e-05, "loss": 1.7753, "step": 1260 }, { "epoch": 0.13628012536474657, "grad_norm": 0.3723194897174835, "learning_rate": 8e-05, "loss": 1.8516, "step": 1261 }, { "epoch": 0.13638819842213337, "grad_norm": 0.34464162588119507, "learning_rate": 8e-05, "loss": 1.7818, "step": 1262 }, { "epoch": 0.13649627147952015, "grad_norm": 0.379607230424881, "learning_rate": 8e-05, "loss": 2.0066, "step": 1263 }, { "epoch": 0.13660434453690695, "grad_norm": 0.35011014342308044, "learning_rate": 8e-05, "loss": 1.7616, "step": 1264 }, { "epoch": 0.13671241759429376, "grad_norm": 0.40158697962760925, "learning_rate": 8e-05, "loss": 1.9133, "step": 1265 }, { "epoch": 0.13682049065168053, "grad_norm": 0.3566371500492096, "learning_rate": 8e-05, "loss": 1.7306, "step": 1266 }, { "epoch": 0.13692856370906734, "grad_norm": 0.3662523925304413, "learning_rate": 8e-05, "loss": 1.8725, "step": 1267 }, { "epoch": 0.1370366367664541, "grad_norm": 0.36400070786476135, "learning_rate": 8e-05, "loss": 1.8992, "step": 1268 }, { "epoch": 0.13714470982384092, "grad_norm": 0.3628464341163635, "learning_rate": 8e-05, "loss": 1.8345, "step": 1269 }, { "epoch": 0.13725278288122772, "grad_norm": 0.3793184161186218, "learning_rate": 8e-05, "loss": 1.8457, "step": 1270 }, { "epoch": 0.1373608559386145, "grad_norm": 0.34508460760116577, "learning_rate": 8e-05, "loss": 1.6921, "step": 1271 }, { "epoch": 0.1374689289960013, "grad_norm": 0.37388306856155396, "learning_rate": 8e-05, "loss": 1.9294, "step": 1272 }, { "epoch": 0.1375770020533881, "grad_norm": 0.37191563844680786, "learning_rate": 8e-05, "loss": 1.6548, "step": 1273 }, { "epoch": 0.13768507511077488, "grad_norm": 0.38890841603279114, "learning_rate": 8e-05, "loss": 1.9279, "step": 1274 }, { "epoch": 0.13779314816816168, "grad_norm": 0.35622289776802063, "learning_rate": 8e-05, "loss": 1.7508, "step": 1275 }, { "epoch": 0.13790122122554846, "grad_norm": 0.3675192594528198, "learning_rate": 8e-05, "loss": 1.6861, "step": 1276 }, { "epoch": 0.13800929428293526, "grad_norm": 0.35501307249069214, "learning_rate": 8e-05, "loss": 1.8708, "step": 1277 }, { "epoch": 0.13811736734032207, "grad_norm": 0.33192023634910583, "learning_rate": 8e-05, "loss": 1.741, "step": 1278 }, { "epoch": 0.13822544039770884, "grad_norm": 0.3726126253604889, "learning_rate": 8e-05, "loss": 1.946, "step": 1279 }, { "epoch": 0.13833351345509565, "grad_norm": 0.34756410121917725, "learning_rate": 8e-05, "loss": 1.7501, "step": 1280 }, { "epoch": 0.13844158651248245, "grad_norm": 0.3411407768726349, "learning_rate": 8e-05, "loss": 1.6545, "step": 1281 }, { "epoch": 0.13854965956986923, "grad_norm": 0.34891775250434875, "learning_rate": 8e-05, "loss": 1.7104, "step": 1282 }, { "epoch": 0.13865773262725603, "grad_norm": 0.36273065209388733, "learning_rate": 8e-05, "loss": 1.7224, "step": 1283 }, { "epoch": 0.1387658056846428, "grad_norm": 0.34276318550109863, "learning_rate": 8e-05, "loss": 1.7133, "step": 1284 }, { "epoch": 0.1388738787420296, "grad_norm": 0.3624266982078552, "learning_rate": 8e-05, "loss": 1.8052, "step": 1285 }, { "epoch": 0.13898195179941641, "grad_norm": 0.3275555968284607, "learning_rate": 8e-05, "loss": 1.6077, "step": 1286 }, { "epoch": 0.1390900248568032, "grad_norm": 0.34999197721481323, "learning_rate": 8e-05, "loss": 1.7905, "step": 1287 }, { "epoch": 0.13919809791419, "grad_norm": 0.3496024012565613, "learning_rate": 8e-05, "loss": 1.7994, "step": 1288 }, { "epoch": 0.1393061709715768, "grad_norm": 0.361912339925766, "learning_rate": 8e-05, "loss": 1.772, "step": 1289 }, { "epoch": 0.13941424402896357, "grad_norm": 0.3756738603115082, "learning_rate": 8e-05, "loss": 1.9363, "step": 1290 }, { "epoch": 0.13952231708635038, "grad_norm": 0.3466061055660248, "learning_rate": 8e-05, "loss": 1.5753, "step": 1291 }, { "epoch": 0.13963039014373715, "grad_norm": 0.36808082461357117, "learning_rate": 8e-05, "loss": 1.8752, "step": 1292 }, { "epoch": 0.13973846320112396, "grad_norm": 0.3684549927711487, "learning_rate": 8e-05, "loss": 1.7835, "step": 1293 }, { "epoch": 0.13984653625851076, "grad_norm": 0.36353185772895813, "learning_rate": 8e-05, "loss": 1.8672, "step": 1294 }, { "epoch": 0.13995460931589754, "grad_norm": 0.3836422860622406, "learning_rate": 8e-05, "loss": 1.8402, "step": 1295 }, { "epoch": 0.14006268237328434, "grad_norm": 0.3743646740913391, "learning_rate": 8e-05, "loss": 1.9166, "step": 1296 }, { "epoch": 0.14017075543067115, "grad_norm": 0.3598948121070862, "learning_rate": 8e-05, "loss": 1.8189, "step": 1297 }, { "epoch": 0.14027882848805792, "grad_norm": 0.38350221514701843, "learning_rate": 8e-05, "loss": 1.8866, "step": 1298 }, { "epoch": 0.14038690154544473, "grad_norm": 0.36876654624938965, "learning_rate": 8e-05, "loss": 1.8262, "step": 1299 }, { "epoch": 0.1404949746028315, "grad_norm": 0.38774171471595764, "learning_rate": 8e-05, "loss": 1.8222, "step": 1300 }, { "epoch": 0.1406030476602183, "grad_norm": 0.37729960680007935, "learning_rate": 8e-05, "loss": 1.673, "step": 1301 }, { "epoch": 0.1407111207176051, "grad_norm": 0.33797571063041687, "learning_rate": 8e-05, "loss": 1.6448, "step": 1302 }, { "epoch": 0.14081919377499189, "grad_norm": 0.34323862195014954, "learning_rate": 8e-05, "loss": 1.7167, "step": 1303 }, { "epoch": 0.1409272668323787, "grad_norm": 0.3635108470916748, "learning_rate": 8e-05, "loss": 1.9497, "step": 1304 }, { "epoch": 0.1410353398897655, "grad_norm": 0.38030102849006653, "learning_rate": 8e-05, "loss": 1.7682, "step": 1305 }, { "epoch": 0.14114341294715227, "grad_norm": 0.3572438955307007, "learning_rate": 8e-05, "loss": 1.7784, "step": 1306 }, { "epoch": 0.14125148600453907, "grad_norm": 0.3755241632461548, "learning_rate": 8e-05, "loss": 1.8321, "step": 1307 }, { "epoch": 0.14135955906192585, "grad_norm": 0.36530229449272156, "learning_rate": 8e-05, "loss": 1.7099, "step": 1308 }, { "epoch": 0.14146763211931265, "grad_norm": 0.38700082898139954, "learning_rate": 8e-05, "loss": 1.8099, "step": 1309 }, { "epoch": 0.14157570517669946, "grad_norm": 0.358844518661499, "learning_rate": 8e-05, "loss": 1.6388, "step": 1310 }, { "epoch": 0.14168377823408623, "grad_norm": 0.3239356577396393, "learning_rate": 8e-05, "loss": 1.7064, "step": 1311 }, { "epoch": 0.14179185129147304, "grad_norm": 0.37981265783309937, "learning_rate": 8e-05, "loss": 1.9305, "step": 1312 }, { "epoch": 0.14189992434885984, "grad_norm": 0.355438232421875, "learning_rate": 8e-05, "loss": 1.7814, "step": 1313 }, { "epoch": 0.14200799740624662, "grad_norm": 0.38024845719337463, "learning_rate": 8e-05, "loss": 1.786, "step": 1314 }, { "epoch": 0.14211607046363342, "grad_norm": 0.33766257762908936, "learning_rate": 8e-05, "loss": 1.6762, "step": 1315 }, { "epoch": 0.1422241435210202, "grad_norm": 0.35067158937454224, "learning_rate": 8e-05, "loss": 1.7361, "step": 1316 }, { "epoch": 0.142332216578407, "grad_norm": 0.36893215775489807, "learning_rate": 8e-05, "loss": 1.8138, "step": 1317 }, { "epoch": 0.1424402896357938, "grad_norm": 0.402853399515152, "learning_rate": 8e-05, "loss": 1.9631, "step": 1318 }, { "epoch": 0.14254836269318058, "grad_norm": 0.3573390245437622, "learning_rate": 8e-05, "loss": 1.5288, "step": 1319 }, { "epoch": 0.14265643575056738, "grad_norm": 0.36313819885253906, "learning_rate": 8e-05, "loss": 1.7057, "step": 1320 }, { "epoch": 0.1427645088079542, "grad_norm": 0.4008879065513611, "learning_rate": 8e-05, "loss": 1.8633, "step": 1321 }, { "epoch": 0.14287258186534096, "grad_norm": 0.38593950867652893, "learning_rate": 8e-05, "loss": 1.7766, "step": 1322 }, { "epoch": 0.14298065492272777, "grad_norm": 0.3522358536720276, "learning_rate": 8e-05, "loss": 1.7343, "step": 1323 }, { "epoch": 0.14308872798011454, "grad_norm": 0.34919023513793945, "learning_rate": 8e-05, "loss": 1.7858, "step": 1324 }, { "epoch": 0.14319680103750135, "grad_norm": 0.4303407073020935, "learning_rate": 8e-05, "loss": 1.7033, "step": 1325 }, { "epoch": 0.14330487409488815, "grad_norm": 0.34750887751579285, "learning_rate": 8e-05, "loss": 1.7438, "step": 1326 }, { "epoch": 0.14341294715227493, "grad_norm": 0.34084826707839966, "learning_rate": 8e-05, "loss": 1.6366, "step": 1327 }, { "epoch": 0.14352102020966173, "grad_norm": 0.3843788504600525, "learning_rate": 8e-05, "loss": 1.8638, "step": 1328 }, { "epoch": 0.14362909326704854, "grad_norm": 0.36963194608688354, "learning_rate": 8e-05, "loss": 1.817, "step": 1329 }, { "epoch": 0.1437371663244353, "grad_norm": 0.3810136318206787, "learning_rate": 8e-05, "loss": 1.7495, "step": 1330 }, { "epoch": 0.14384523938182212, "grad_norm": 0.36227381229400635, "learning_rate": 8e-05, "loss": 1.8458, "step": 1331 }, { "epoch": 0.1439533124392089, "grad_norm": 0.3583066165447235, "learning_rate": 8e-05, "loss": 1.8114, "step": 1332 }, { "epoch": 0.1440613854965957, "grad_norm": 0.3827393352985382, "learning_rate": 8e-05, "loss": 1.8215, "step": 1333 }, { "epoch": 0.1441694585539825, "grad_norm": 0.38677769899368286, "learning_rate": 8e-05, "loss": 1.8207, "step": 1334 }, { "epoch": 0.14427753161136928, "grad_norm": 0.35298940539360046, "learning_rate": 8e-05, "loss": 1.7273, "step": 1335 }, { "epoch": 0.14438560466875608, "grad_norm": 0.353473961353302, "learning_rate": 8e-05, "loss": 1.7931, "step": 1336 }, { "epoch": 0.14449367772614288, "grad_norm": 0.3717522621154785, "learning_rate": 8e-05, "loss": 1.9448, "step": 1337 }, { "epoch": 0.14460175078352966, "grad_norm": 0.36258256435394287, "learning_rate": 8e-05, "loss": 1.794, "step": 1338 }, { "epoch": 0.14470982384091646, "grad_norm": 0.3871307075023651, "learning_rate": 8e-05, "loss": 1.8851, "step": 1339 }, { "epoch": 0.14481789689830327, "grad_norm": 0.3597882091999054, "learning_rate": 8e-05, "loss": 1.7949, "step": 1340 }, { "epoch": 0.14492596995569004, "grad_norm": 0.3686104416847229, "learning_rate": 8e-05, "loss": 1.786, "step": 1341 }, { "epoch": 0.14503404301307685, "grad_norm": 0.38345858454704285, "learning_rate": 8e-05, "loss": 1.9195, "step": 1342 }, { "epoch": 0.14514211607046362, "grad_norm": 0.3607224225997925, "learning_rate": 8e-05, "loss": 1.9278, "step": 1343 }, { "epoch": 0.14525018912785043, "grad_norm": 0.33355677127838135, "learning_rate": 8e-05, "loss": 1.7393, "step": 1344 }, { "epoch": 0.14535826218523723, "grad_norm": 0.35089266300201416, "learning_rate": 8e-05, "loss": 1.674, "step": 1345 }, { "epoch": 0.145466335242624, "grad_norm": 0.35067078471183777, "learning_rate": 8e-05, "loss": 1.6902, "step": 1346 }, { "epoch": 0.1455744083000108, "grad_norm": 0.35832321643829346, "learning_rate": 8e-05, "loss": 1.7072, "step": 1347 }, { "epoch": 0.14568248135739761, "grad_norm": 0.36257049441337585, "learning_rate": 8e-05, "loss": 1.7685, "step": 1348 }, { "epoch": 0.1457905544147844, "grad_norm": 0.36324799060821533, "learning_rate": 8e-05, "loss": 1.792, "step": 1349 }, { "epoch": 0.1458986274721712, "grad_norm": 0.37073221802711487, "learning_rate": 8e-05, "loss": 1.7197, "step": 1350 }, { "epoch": 0.14600670052955797, "grad_norm": 0.39439767599105835, "learning_rate": 8e-05, "loss": 1.638, "step": 1351 }, { "epoch": 0.14611477358694477, "grad_norm": 0.428548663854599, "learning_rate": 8e-05, "loss": 1.5965, "step": 1352 }, { "epoch": 0.14622284664433158, "grad_norm": 0.41466617584228516, "learning_rate": 8e-05, "loss": 1.8584, "step": 1353 }, { "epoch": 0.14633091970171835, "grad_norm": 0.3542669415473938, "learning_rate": 8e-05, "loss": 1.7357, "step": 1354 }, { "epoch": 0.14643899275910516, "grad_norm": 0.34826070070266724, "learning_rate": 8e-05, "loss": 1.7364, "step": 1355 }, { "epoch": 0.14654706581649196, "grad_norm": 0.3319184482097626, "learning_rate": 8e-05, "loss": 1.5644, "step": 1356 }, { "epoch": 0.14665513887387874, "grad_norm": 0.37194564938545227, "learning_rate": 8e-05, "loss": 1.854, "step": 1357 }, { "epoch": 0.14676321193126554, "grad_norm": 0.37797489762306213, "learning_rate": 8e-05, "loss": 1.6787, "step": 1358 }, { "epoch": 0.14687128498865232, "grad_norm": 0.3618971109390259, "learning_rate": 8e-05, "loss": 1.8924, "step": 1359 }, { "epoch": 0.14697935804603912, "grad_norm": 0.38728848099708557, "learning_rate": 8e-05, "loss": 1.8663, "step": 1360 }, { "epoch": 0.14708743110342593, "grad_norm": 0.36954250931739807, "learning_rate": 8e-05, "loss": 1.836, "step": 1361 }, { "epoch": 0.1471955041608127, "grad_norm": 0.3604555130004883, "learning_rate": 8e-05, "loss": 1.7374, "step": 1362 }, { "epoch": 0.1473035772181995, "grad_norm": 0.3396654725074768, "learning_rate": 8e-05, "loss": 1.6719, "step": 1363 }, { "epoch": 0.1474116502755863, "grad_norm": 0.3587683141231537, "learning_rate": 8e-05, "loss": 1.7432, "step": 1364 }, { "epoch": 0.14751972333297309, "grad_norm": 0.36430948972702026, "learning_rate": 8e-05, "loss": 1.7602, "step": 1365 }, { "epoch": 0.1476277963903599, "grad_norm": 0.3922954201698303, "learning_rate": 8e-05, "loss": 1.8045, "step": 1366 }, { "epoch": 0.14773586944774666, "grad_norm": 0.3604581952095032, "learning_rate": 8e-05, "loss": 1.8026, "step": 1367 }, { "epoch": 0.14784394250513347, "grad_norm": 0.35327261686325073, "learning_rate": 8e-05, "loss": 1.7987, "step": 1368 }, { "epoch": 0.14795201556252027, "grad_norm": 0.3642292022705078, "learning_rate": 8e-05, "loss": 1.8837, "step": 1369 }, { "epoch": 0.14806008861990705, "grad_norm": 0.3424132764339447, "learning_rate": 8e-05, "loss": 1.698, "step": 1370 }, { "epoch": 0.14816816167729385, "grad_norm": 0.36793118715286255, "learning_rate": 8e-05, "loss": 1.7402, "step": 1371 }, { "epoch": 0.14827623473468066, "grad_norm": 0.37847766280174255, "learning_rate": 8e-05, "loss": 1.9174, "step": 1372 }, { "epoch": 0.14838430779206743, "grad_norm": 0.3926122188568115, "learning_rate": 8e-05, "loss": 2.0138, "step": 1373 }, { "epoch": 0.14849238084945424, "grad_norm": 0.3614753484725952, "learning_rate": 8e-05, "loss": 1.8223, "step": 1374 }, { "epoch": 0.148600453906841, "grad_norm": 0.3655136525630951, "learning_rate": 8e-05, "loss": 1.6286, "step": 1375 }, { "epoch": 0.14870852696422782, "grad_norm": 0.36882254481315613, "learning_rate": 8e-05, "loss": 1.7201, "step": 1376 }, { "epoch": 0.14881660002161462, "grad_norm": 0.3873593509197235, "learning_rate": 8e-05, "loss": 1.9422, "step": 1377 }, { "epoch": 0.1489246730790014, "grad_norm": 0.38361072540283203, "learning_rate": 8e-05, "loss": 1.7471, "step": 1378 }, { "epoch": 0.1490327461363882, "grad_norm": 0.34324052929878235, "learning_rate": 8e-05, "loss": 1.6718, "step": 1379 }, { "epoch": 0.149140819193775, "grad_norm": 0.3797532618045807, "learning_rate": 8e-05, "loss": 1.9269, "step": 1380 }, { "epoch": 0.14924889225116178, "grad_norm": 0.40686362981796265, "learning_rate": 8e-05, "loss": 1.9187, "step": 1381 }, { "epoch": 0.14935696530854858, "grad_norm": 0.35567137598991394, "learning_rate": 8e-05, "loss": 1.8577, "step": 1382 }, { "epoch": 0.14946503836593536, "grad_norm": 0.3699253499507904, "learning_rate": 8e-05, "loss": 1.8638, "step": 1383 }, { "epoch": 0.14957311142332216, "grad_norm": 0.3547039031982422, "learning_rate": 8e-05, "loss": 1.7259, "step": 1384 }, { "epoch": 0.14968118448070897, "grad_norm": 0.3718073070049286, "learning_rate": 8e-05, "loss": 1.7663, "step": 1385 }, { "epoch": 0.14978925753809574, "grad_norm": 0.37713155150413513, "learning_rate": 8e-05, "loss": 1.8828, "step": 1386 }, { "epoch": 0.14989733059548255, "grad_norm": 0.36620232462882996, "learning_rate": 8e-05, "loss": 1.7877, "step": 1387 }, { "epoch": 0.15000540365286935, "grad_norm": 0.34448128938674927, "learning_rate": 8e-05, "loss": 1.5905, "step": 1388 }, { "epoch": 0.15011347671025613, "grad_norm": 0.37907782196998596, "learning_rate": 8e-05, "loss": 1.5522, "step": 1389 }, { "epoch": 0.15022154976764293, "grad_norm": 0.3553794324398041, "learning_rate": 8e-05, "loss": 1.8251, "step": 1390 }, { "epoch": 0.1503296228250297, "grad_norm": 0.3589995503425598, "learning_rate": 8e-05, "loss": 1.807, "step": 1391 }, { "epoch": 0.1504376958824165, "grad_norm": 0.3628771901130676, "learning_rate": 8e-05, "loss": 1.7974, "step": 1392 }, { "epoch": 0.15054576893980331, "grad_norm": 0.36628058552742004, "learning_rate": 8e-05, "loss": 1.7988, "step": 1393 }, { "epoch": 0.1506538419971901, "grad_norm": 0.35752347111701965, "learning_rate": 8e-05, "loss": 1.8529, "step": 1394 }, { "epoch": 0.1507619150545769, "grad_norm": 0.3432478606700897, "learning_rate": 8e-05, "loss": 1.6778, "step": 1395 }, { "epoch": 0.1508699881119637, "grad_norm": 0.33416950702667236, "learning_rate": 8e-05, "loss": 1.6821, "step": 1396 }, { "epoch": 0.15097806116935047, "grad_norm": 0.3986573815345764, "learning_rate": 8e-05, "loss": 1.9902, "step": 1397 }, { "epoch": 0.15108613422673728, "grad_norm": 0.36368662118911743, "learning_rate": 8e-05, "loss": 1.8914, "step": 1398 }, { "epoch": 0.15119420728412405, "grad_norm": 0.39978787302970886, "learning_rate": 8e-05, "loss": 1.8257, "step": 1399 }, { "epoch": 0.15130228034151086, "grad_norm": 0.35373568534851074, "learning_rate": 8e-05, "loss": 1.7018, "step": 1400 }, { "epoch": 0.15141035339889766, "grad_norm": 0.3432331085205078, "learning_rate": 8e-05, "loss": 1.7616, "step": 1401 }, { "epoch": 0.15151842645628444, "grad_norm": 0.36555132269859314, "learning_rate": 8e-05, "loss": 1.8551, "step": 1402 }, { "epoch": 0.15162649951367124, "grad_norm": 0.3434751629829407, "learning_rate": 8e-05, "loss": 1.6457, "step": 1403 }, { "epoch": 0.15173457257105805, "grad_norm": 0.3762097656726837, "learning_rate": 8e-05, "loss": 1.8822, "step": 1404 }, { "epoch": 0.15184264562844482, "grad_norm": 0.3662976026535034, "learning_rate": 8e-05, "loss": 1.7431, "step": 1405 }, { "epoch": 0.15195071868583163, "grad_norm": 0.3371051251888275, "learning_rate": 8e-05, "loss": 1.6064, "step": 1406 }, { "epoch": 0.1520587917432184, "grad_norm": 0.36757299304008484, "learning_rate": 8e-05, "loss": 1.7929, "step": 1407 }, { "epoch": 0.1521668648006052, "grad_norm": 0.3737618029117584, "learning_rate": 8e-05, "loss": 1.8545, "step": 1408 }, { "epoch": 0.152274937857992, "grad_norm": 0.4045862853527069, "learning_rate": 8e-05, "loss": 1.9242, "step": 1409 }, { "epoch": 0.15238301091537879, "grad_norm": 0.4126172959804535, "learning_rate": 8e-05, "loss": 1.9788, "step": 1410 }, { "epoch": 0.1524910839727656, "grad_norm": 0.3461454510688782, "learning_rate": 8e-05, "loss": 1.7328, "step": 1411 }, { "epoch": 0.1525991570301524, "grad_norm": 0.3536359965801239, "learning_rate": 8e-05, "loss": 1.8372, "step": 1412 }, { "epoch": 0.15270723008753917, "grad_norm": 0.37129977345466614, "learning_rate": 8e-05, "loss": 1.828, "step": 1413 }, { "epoch": 0.15281530314492597, "grad_norm": 0.3773781359195709, "learning_rate": 8e-05, "loss": 1.9379, "step": 1414 }, { "epoch": 0.15292337620231275, "grad_norm": 0.3861599266529083, "learning_rate": 8e-05, "loss": 1.8157, "step": 1415 }, { "epoch": 0.15303144925969955, "grad_norm": 0.3463456332683563, "learning_rate": 8e-05, "loss": 1.793, "step": 1416 }, { "epoch": 0.15313952231708636, "grad_norm": 0.363972932100296, "learning_rate": 8e-05, "loss": 1.7568, "step": 1417 }, { "epoch": 0.15324759537447313, "grad_norm": 0.38853877782821655, "learning_rate": 8e-05, "loss": 1.9306, "step": 1418 }, { "epoch": 0.15335566843185994, "grad_norm": 0.3726254105567932, "learning_rate": 8e-05, "loss": 1.8972, "step": 1419 }, { "epoch": 0.15346374148924674, "grad_norm": 0.3519631326198578, "learning_rate": 8e-05, "loss": 1.7771, "step": 1420 }, { "epoch": 0.15357181454663352, "grad_norm": 0.3772374093532562, "learning_rate": 8e-05, "loss": 1.7649, "step": 1421 }, { "epoch": 0.15367988760402032, "grad_norm": 0.34010839462280273, "learning_rate": 8e-05, "loss": 1.7287, "step": 1422 }, { "epoch": 0.15378796066140712, "grad_norm": 0.3800028860569, "learning_rate": 8e-05, "loss": 1.7983, "step": 1423 }, { "epoch": 0.1538960337187939, "grad_norm": 0.34591197967529297, "learning_rate": 8e-05, "loss": 1.7964, "step": 1424 }, { "epoch": 0.1540041067761807, "grad_norm": 0.35065922141075134, "learning_rate": 8e-05, "loss": 1.5771, "step": 1425 }, { "epoch": 0.15411217983356748, "grad_norm": 0.3294548988342285, "learning_rate": 8e-05, "loss": 1.6164, "step": 1426 }, { "epoch": 0.15422025289095428, "grad_norm": 0.35767996311187744, "learning_rate": 8e-05, "loss": 1.7424, "step": 1427 }, { "epoch": 0.1543283259483411, "grad_norm": 0.4050959348678589, "learning_rate": 8e-05, "loss": 1.9109, "step": 1428 }, { "epoch": 0.15443639900572786, "grad_norm": 0.38854584097862244, "learning_rate": 8e-05, "loss": 1.8469, "step": 1429 }, { "epoch": 0.15454447206311467, "grad_norm": 0.3645983338356018, "learning_rate": 8e-05, "loss": 1.7182, "step": 1430 }, { "epoch": 0.15465254512050147, "grad_norm": 0.3580503761768341, "learning_rate": 8e-05, "loss": 1.7341, "step": 1431 }, { "epoch": 0.15476061817788825, "grad_norm": 0.38257795572280884, "learning_rate": 8e-05, "loss": 1.5142, "step": 1432 }, { "epoch": 0.15486869123527505, "grad_norm": 0.3839014172554016, "learning_rate": 8e-05, "loss": 1.8363, "step": 1433 }, { "epoch": 0.15497676429266183, "grad_norm": 0.3898155987262726, "learning_rate": 8e-05, "loss": 1.8737, "step": 1434 }, { "epoch": 0.15508483735004863, "grad_norm": 0.3901798129081726, "learning_rate": 8e-05, "loss": 1.9292, "step": 1435 }, { "epoch": 0.15519291040743544, "grad_norm": 0.3728411793708801, "learning_rate": 8e-05, "loss": 1.9086, "step": 1436 }, { "epoch": 0.1553009834648222, "grad_norm": 0.32981374859809875, "learning_rate": 8e-05, "loss": 1.7063, "step": 1437 }, { "epoch": 0.15540905652220902, "grad_norm": 0.37500622868537903, "learning_rate": 8e-05, "loss": 2.0201, "step": 1438 }, { "epoch": 0.15551712957959582, "grad_norm": 0.36162200570106506, "learning_rate": 8e-05, "loss": 1.6942, "step": 1439 }, { "epoch": 0.1556252026369826, "grad_norm": 0.35528993606567383, "learning_rate": 8e-05, "loss": 1.6467, "step": 1440 }, { "epoch": 0.1557332756943694, "grad_norm": 0.3630034625530243, "learning_rate": 8e-05, "loss": 1.8, "step": 1441 }, { "epoch": 0.15584134875175618, "grad_norm": 0.3964095115661621, "learning_rate": 8e-05, "loss": 1.7402, "step": 1442 }, { "epoch": 0.15594942180914298, "grad_norm": 0.3552529811859131, "learning_rate": 8e-05, "loss": 1.6099, "step": 1443 }, { "epoch": 0.15605749486652978, "grad_norm": 0.3829663097858429, "learning_rate": 8e-05, "loss": 1.7102, "step": 1444 }, { "epoch": 0.15616556792391656, "grad_norm": 0.35981523990631104, "learning_rate": 8e-05, "loss": 1.7307, "step": 1445 }, { "epoch": 0.15627364098130336, "grad_norm": 0.3516583740711212, "learning_rate": 8e-05, "loss": 1.6101, "step": 1446 }, { "epoch": 0.15638171403869017, "grad_norm": 0.34881412982940674, "learning_rate": 8e-05, "loss": 1.7242, "step": 1447 }, { "epoch": 0.15648978709607694, "grad_norm": 0.40145429968833923, "learning_rate": 8e-05, "loss": 1.7185, "step": 1448 }, { "epoch": 0.15659786015346375, "grad_norm": 0.35522493720054626, "learning_rate": 8e-05, "loss": 1.737, "step": 1449 }, { "epoch": 0.15670593321085052, "grad_norm": 0.3860670328140259, "learning_rate": 8e-05, "loss": 1.8801, "step": 1450 }, { "epoch": 0.15681400626823733, "grad_norm": 0.3494183421134949, "learning_rate": 8e-05, "loss": 1.7717, "step": 1451 }, { "epoch": 0.15692207932562413, "grad_norm": 0.3855258822441101, "learning_rate": 8e-05, "loss": 1.8964, "step": 1452 }, { "epoch": 0.1570301523830109, "grad_norm": 0.3451821804046631, "learning_rate": 8e-05, "loss": 1.6332, "step": 1453 }, { "epoch": 0.1571382254403977, "grad_norm": 0.34593403339385986, "learning_rate": 8e-05, "loss": 1.6426, "step": 1454 }, { "epoch": 0.15724629849778451, "grad_norm": 0.36042019724845886, "learning_rate": 8e-05, "loss": 1.7771, "step": 1455 }, { "epoch": 0.1573543715551713, "grad_norm": 0.3821708559989929, "learning_rate": 8e-05, "loss": 1.8209, "step": 1456 }, { "epoch": 0.1574624446125581, "grad_norm": 0.3659166693687439, "learning_rate": 8e-05, "loss": 1.9368, "step": 1457 }, { "epoch": 0.15757051766994487, "grad_norm": 0.382966548204422, "learning_rate": 8e-05, "loss": 1.9203, "step": 1458 }, { "epoch": 0.15767859072733167, "grad_norm": 0.35953521728515625, "learning_rate": 8e-05, "loss": 1.7595, "step": 1459 }, { "epoch": 0.15778666378471848, "grad_norm": 0.39093002676963806, "learning_rate": 8e-05, "loss": 1.8129, "step": 1460 }, { "epoch": 0.15789473684210525, "grad_norm": 0.3680100739002228, "learning_rate": 8e-05, "loss": 1.8641, "step": 1461 }, { "epoch": 0.15800280989949206, "grad_norm": 0.37109389901161194, "learning_rate": 8e-05, "loss": 1.7952, "step": 1462 }, { "epoch": 0.15811088295687886, "grad_norm": 0.35500001907348633, "learning_rate": 8e-05, "loss": 1.8556, "step": 1463 }, { "epoch": 0.15821895601426564, "grad_norm": 0.3553858995437622, "learning_rate": 8e-05, "loss": 1.7524, "step": 1464 }, { "epoch": 0.15832702907165244, "grad_norm": 0.380461722612381, "learning_rate": 8e-05, "loss": 1.7321, "step": 1465 }, { "epoch": 0.15843510212903922, "grad_norm": 0.36477822065353394, "learning_rate": 8e-05, "loss": 1.7948, "step": 1466 }, { "epoch": 0.15854317518642602, "grad_norm": 0.359115332365036, "learning_rate": 8e-05, "loss": 1.6428, "step": 1467 }, { "epoch": 0.15865124824381283, "grad_norm": 0.3710903823375702, "learning_rate": 8e-05, "loss": 1.8335, "step": 1468 }, { "epoch": 0.1587593213011996, "grad_norm": 0.3482389748096466, "learning_rate": 8e-05, "loss": 1.7727, "step": 1469 }, { "epoch": 0.1588673943585864, "grad_norm": 0.3669455647468567, "learning_rate": 8e-05, "loss": 1.7746, "step": 1470 }, { "epoch": 0.1589754674159732, "grad_norm": 0.37701231241226196, "learning_rate": 8e-05, "loss": 1.7545, "step": 1471 }, { "epoch": 0.15908354047335999, "grad_norm": 0.3684040307998657, "learning_rate": 8e-05, "loss": 1.7844, "step": 1472 }, { "epoch": 0.1591916135307468, "grad_norm": 0.3648243248462677, "learning_rate": 8e-05, "loss": 1.6594, "step": 1473 }, { "epoch": 0.15929968658813357, "grad_norm": 0.3526931703090668, "learning_rate": 8e-05, "loss": 1.7759, "step": 1474 }, { "epoch": 0.15940775964552037, "grad_norm": 0.3462124466896057, "learning_rate": 8e-05, "loss": 1.6768, "step": 1475 }, { "epoch": 0.15951583270290717, "grad_norm": 0.37210389971733093, "learning_rate": 8e-05, "loss": 1.8001, "step": 1476 }, { "epoch": 0.15962390576029395, "grad_norm": 0.3444695472717285, "learning_rate": 8e-05, "loss": 1.6412, "step": 1477 }, { "epoch": 0.15973197881768075, "grad_norm": 0.3394198715686798, "learning_rate": 8e-05, "loss": 1.6644, "step": 1478 }, { "epoch": 0.15984005187506756, "grad_norm": 0.35910144448280334, "learning_rate": 8e-05, "loss": 1.7493, "step": 1479 }, { "epoch": 0.15994812493245433, "grad_norm": 0.34826552867889404, "learning_rate": 8e-05, "loss": 1.8196, "step": 1480 }, { "epoch": 0.16005619798984114, "grad_norm": 0.35231512784957886, "learning_rate": 8e-05, "loss": 1.8121, "step": 1481 }, { "epoch": 0.1601642710472279, "grad_norm": 0.3483321964740753, "learning_rate": 8e-05, "loss": 1.7305, "step": 1482 }, { "epoch": 0.16027234410461472, "grad_norm": 0.3588305711746216, "learning_rate": 8e-05, "loss": 1.7277, "step": 1483 }, { "epoch": 0.16038041716200152, "grad_norm": 0.3416995406150818, "learning_rate": 8e-05, "loss": 1.6039, "step": 1484 }, { "epoch": 0.1604884902193883, "grad_norm": 0.3925912082195282, "learning_rate": 8e-05, "loss": 1.9478, "step": 1485 }, { "epoch": 0.1605965632767751, "grad_norm": 0.34834226965904236, "learning_rate": 8e-05, "loss": 1.7404, "step": 1486 }, { "epoch": 0.1607046363341619, "grad_norm": 0.36548808217048645, "learning_rate": 8e-05, "loss": 1.6192, "step": 1487 }, { "epoch": 0.16081270939154868, "grad_norm": 0.36002662777900696, "learning_rate": 8e-05, "loss": 1.7988, "step": 1488 }, { "epoch": 0.16092078244893548, "grad_norm": 0.38215070962905884, "learning_rate": 8e-05, "loss": 1.9371, "step": 1489 }, { "epoch": 0.16102885550632226, "grad_norm": 0.3457402288913727, "learning_rate": 8e-05, "loss": 1.6629, "step": 1490 }, { "epoch": 0.16113692856370906, "grad_norm": 0.34562206268310547, "learning_rate": 8e-05, "loss": 1.6431, "step": 1491 }, { "epoch": 0.16124500162109587, "grad_norm": 0.35419896245002747, "learning_rate": 8e-05, "loss": 1.7419, "step": 1492 }, { "epoch": 0.16135307467848264, "grad_norm": 0.339605450630188, "learning_rate": 8e-05, "loss": 1.723, "step": 1493 }, { "epoch": 0.16146114773586945, "grad_norm": 0.35562995076179504, "learning_rate": 8e-05, "loss": 1.8317, "step": 1494 }, { "epoch": 0.16156922079325625, "grad_norm": 0.38061460852622986, "learning_rate": 8e-05, "loss": 1.7959, "step": 1495 }, { "epoch": 0.16167729385064303, "grad_norm": 0.35861071944236755, "learning_rate": 8e-05, "loss": 1.6088, "step": 1496 }, { "epoch": 0.16178536690802983, "grad_norm": 0.42980650067329407, "learning_rate": 8e-05, "loss": 2.0117, "step": 1497 }, { "epoch": 0.1618934399654166, "grad_norm": 0.36173132061958313, "learning_rate": 8e-05, "loss": 1.8477, "step": 1498 }, { "epoch": 0.1620015130228034, "grad_norm": 0.35599803924560547, "learning_rate": 8e-05, "loss": 1.7989, "step": 1499 }, { "epoch": 0.16210958608019022, "grad_norm": 0.3865572214126587, "learning_rate": 8e-05, "loss": 1.6345, "step": 1500 }, { "epoch": 0.162217659137577, "grad_norm": 0.38279634714126587, "learning_rate": 8e-05, "loss": 1.8209, "step": 1501 }, { "epoch": 0.1623257321949638, "grad_norm": 0.35367077589035034, "learning_rate": 8e-05, "loss": 1.7587, "step": 1502 }, { "epoch": 0.1624338052523506, "grad_norm": 0.36740198731422424, "learning_rate": 8e-05, "loss": 1.7575, "step": 1503 }, { "epoch": 0.16254187830973738, "grad_norm": 0.3541874289512634, "learning_rate": 8e-05, "loss": 1.7103, "step": 1504 }, { "epoch": 0.16264995136712418, "grad_norm": 0.35867244005203247, "learning_rate": 8e-05, "loss": 1.7863, "step": 1505 }, { "epoch": 0.16275802442451098, "grad_norm": 0.3447834849357605, "learning_rate": 8e-05, "loss": 1.6629, "step": 1506 }, { "epoch": 0.16286609748189776, "grad_norm": 0.353619247674942, "learning_rate": 8e-05, "loss": 1.5825, "step": 1507 }, { "epoch": 0.16297417053928456, "grad_norm": 0.34993064403533936, "learning_rate": 8e-05, "loss": 1.6087, "step": 1508 }, { "epoch": 0.16308224359667134, "grad_norm": 0.3920869529247284, "learning_rate": 8e-05, "loss": 1.8182, "step": 1509 }, { "epoch": 0.16319031665405814, "grad_norm": 0.3991021513938904, "learning_rate": 8e-05, "loss": 1.9579, "step": 1510 }, { "epoch": 0.16329838971144495, "grad_norm": 0.3829973638057709, "learning_rate": 8e-05, "loss": 1.9362, "step": 1511 }, { "epoch": 0.16340646276883172, "grad_norm": 0.3766539692878723, "learning_rate": 8e-05, "loss": 1.7401, "step": 1512 }, { "epoch": 0.16351453582621853, "grad_norm": 0.34730976819992065, "learning_rate": 8e-05, "loss": 1.6612, "step": 1513 }, { "epoch": 0.16362260888360533, "grad_norm": 0.3472457826137543, "learning_rate": 8e-05, "loss": 1.596, "step": 1514 }, { "epoch": 0.1637306819409921, "grad_norm": 0.3421221673488617, "learning_rate": 8e-05, "loss": 1.6434, "step": 1515 }, { "epoch": 0.1638387549983789, "grad_norm": 0.3415828347206116, "learning_rate": 8e-05, "loss": 1.6998, "step": 1516 }, { "epoch": 0.1639468280557657, "grad_norm": 0.35462209582328796, "learning_rate": 8e-05, "loss": 1.6565, "step": 1517 }, { "epoch": 0.1640549011131525, "grad_norm": 0.34254932403564453, "learning_rate": 8e-05, "loss": 1.72, "step": 1518 }, { "epoch": 0.1641629741705393, "grad_norm": 0.3488869071006775, "learning_rate": 8e-05, "loss": 1.7057, "step": 1519 }, { "epoch": 0.16427104722792607, "grad_norm": 0.39570122957229614, "learning_rate": 8e-05, "loss": 1.7605, "step": 1520 }, { "epoch": 0.16437912028531287, "grad_norm": 0.3649577498435974, "learning_rate": 8e-05, "loss": 1.8456, "step": 1521 }, { "epoch": 0.16448719334269968, "grad_norm": 0.3431411385536194, "learning_rate": 8e-05, "loss": 1.7864, "step": 1522 }, { "epoch": 0.16459526640008645, "grad_norm": 0.3649054765701294, "learning_rate": 8e-05, "loss": 1.7539, "step": 1523 }, { "epoch": 0.16470333945747326, "grad_norm": 0.39293426275253296, "learning_rate": 8e-05, "loss": 1.8406, "step": 1524 }, { "epoch": 0.16481141251486003, "grad_norm": 0.40931570529937744, "learning_rate": 8e-05, "loss": 1.7165, "step": 1525 }, { "epoch": 0.16491948557224684, "grad_norm": 0.38008323311805725, "learning_rate": 8e-05, "loss": 1.9167, "step": 1526 }, { "epoch": 0.16502755862963364, "grad_norm": 0.4023352563381195, "learning_rate": 8e-05, "loss": 1.926, "step": 1527 }, { "epoch": 0.16513563168702042, "grad_norm": 0.3599163591861725, "learning_rate": 8e-05, "loss": 1.7472, "step": 1528 }, { "epoch": 0.16524370474440722, "grad_norm": 0.39081862568855286, "learning_rate": 8e-05, "loss": 1.7677, "step": 1529 }, { "epoch": 0.16535177780179403, "grad_norm": 0.36268776655197144, "learning_rate": 8e-05, "loss": 1.94, "step": 1530 }, { "epoch": 0.1654598508591808, "grad_norm": 0.3584979772567749, "learning_rate": 8e-05, "loss": 1.7082, "step": 1531 }, { "epoch": 0.1655679239165676, "grad_norm": 0.36618903279304504, "learning_rate": 8e-05, "loss": 1.8008, "step": 1532 }, { "epoch": 0.16567599697395438, "grad_norm": 0.3684219419956207, "learning_rate": 8e-05, "loss": 1.7516, "step": 1533 }, { "epoch": 0.16578407003134119, "grad_norm": 0.3563080132007599, "learning_rate": 8e-05, "loss": 1.7981, "step": 1534 }, { "epoch": 0.165892143088728, "grad_norm": 0.3727073073387146, "learning_rate": 8e-05, "loss": 1.8343, "step": 1535 }, { "epoch": 0.16600021614611477, "grad_norm": 0.3701125681400299, "learning_rate": 8e-05, "loss": 1.7806, "step": 1536 }, { "epoch": 0.16610828920350157, "grad_norm": 0.37388166785240173, "learning_rate": 8e-05, "loss": 1.7219, "step": 1537 }, { "epoch": 0.16621636226088837, "grad_norm": 0.3690321147441864, "learning_rate": 8e-05, "loss": 1.7482, "step": 1538 }, { "epoch": 0.16632443531827515, "grad_norm": 0.36835116147994995, "learning_rate": 8e-05, "loss": 1.7171, "step": 1539 }, { "epoch": 0.16643250837566195, "grad_norm": 0.3663565516471863, "learning_rate": 8e-05, "loss": 1.7528, "step": 1540 }, { "epoch": 0.16654058143304873, "grad_norm": 0.3550228774547577, "learning_rate": 8e-05, "loss": 1.7521, "step": 1541 }, { "epoch": 0.16664865449043553, "grad_norm": 0.35605695843696594, "learning_rate": 8e-05, "loss": 1.6268, "step": 1542 }, { "epoch": 0.16675672754782234, "grad_norm": 0.373093843460083, "learning_rate": 8e-05, "loss": 1.8254, "step": 1543 }, { "epoch": 0.1668648006052091, "grad_norm": 0.36627182364463806, "learning_rate": 8e-05, "loss": 1.5629, "step": 1544 }, { "epoch": 0.16697287366259592, "grad_norm": 0.358193039894104, "learning_rate": 8e-05, "loss": 1.8812, "step": 1545 }, { "epoch": 0.16708094671998272, "grad_norm": 0.38430899381637573, "learning_rate": 8e-05, "loss": 1.7389, "step": 1546 }, { "epoch": 0.1671890197773695, "grad_norm": 0.36285752058029175, "learning_rate": 8e-05, "loss": 1.7497, "step": 1547 }, { "epoch": 0.1672970928347563, "grad_norm": 0.35644498467445374, "learning_rate": 8e-05, "loss": 1.7284, "step": 1548 }, { "epoch": 0.16740516589214308, "grad_norm": 0.37135079503059387, "learning_rate": 8e-05, "loss": 1.7624, "step": 1549 }, { "epoch": 0.16751323894952988, "grad_norm": 0.40812957286834717, "learning_rate": 8e-05, "loss": 1.9663, "step": 1550 }, { "epoch": 0.16762131200691668, "grad_norm": 0.3721383810043335, "learning_rate": 8e-05, "loss": 1.8155, "step": 1551 }, { "epoch": 0.16772938506430346, "grad_norm": 0.343982994556427, "learning_rate": 8e-05, "loss": 1.9217, "step": 1552 }, { "epoch": 0.16783745812169026, "grad_norm": 0.36501482129096985, "learning_rate": 8e-05, "loss": 1.8496, "step": 1553 }, { "epoch": 0.16794553117907707, "grad_norm": 0.36781203746795654, "learning_rate": 8e-05, "loss": 1.7314, "step": 1554 }, { "epoch": 0.16805360423646384, "grad_norm": 0.372563898563385, "learning_rate": 8e-05, "loss": 1.9814, "step": 1555 }, { "epoch": 0.16816167729385065, "grad_norm": 0.38038164377212524, "learning_rate": 8e-05, "loss": 1.8364, "step": 1556 }, { "epoch": 0.16826975035123742, "grad_norm": 0.40095072984695435, "learning_rate": 8e-05, "loss": 1.8653, "step": 1557 }, { "epoch": 0.16837782340862423, "grad_norm": 0.34620606899261475, "learning_rate": 8e-05, "loss": 1.6956, "step": 1558 }, { "epoch": 0.16848589646601103, "grad_norm": 0.37505319714546204, "learning_rate": 8e-05, "loss": 1.6688, "step": 1559 }, { "epoch": 0.1685939695233978, "grad_norm": 0.3739761412143707, "learning_rate": 8e-05, "loss": 1.8345, "step": 1560 }, { "epoch": 0.1687020425807846, "grad_norm": 0.3686574101448059, "learning_rate": 8e-05, "loss": 1.774, "step": 1561 }, { "epoch": 0.16881011563817142, "grad_norm": 0.3540516495704651, "learning_rate": 8e-05, "loss": 1.7203, "step": 1562 }, { "epoch": 0.1689181886955582, "grad_norm": 0.3449227809906006, "learning_rate": 8e-05, "loss": 1.7788, "step": 1563 }, { "epoch": 0.169026261752945, "grad_norm": 0.3845237195491791, "learning_rate": 8e-05, "loss": 1.9325, "step": 1564 }, { "epoch": 0.16913433481033177, "grad_norm": 0.3843061029911041, "learning_rate": 8e-05, "loss": 1.8386, "step": 1565 }, { "epoch": 0.16924240786771858, "grad_norm": 0.4084787964820862, "learning_rate": 8e-05, "loss": 1.9091, "step": 1566 }, { "epoch": 0.16935048092510538, "grad_norm": 0.35452964901924133, "learning_rate": 8e-05, "loss": 1.7746, "step": 1567 }, { "epoch": 0.16945855398249215, "grad_norm": 0.359918475151062, "learning_rate": 8e-05, "loss": 1.881, "step": 1568 }, { "epoch": 0.16956662703987896, "grad_norm": 0.3757287263870239, "learning_rate": 8e-05, "loss": 1.8829, "step": 1569 }, { "epoch": 0.16967470009726576, "grad_norm": 0.37208443880081177, "learning_rate": 8e-05, "loss": 1.7709, "step": 1570 }, { "epoch": 0.16978277315465254, "grad_norm": 0.3668755292892456, "learning_rate": 8e-05, "loss": 1.7275, "step": 1571 }, { "epoch": 0.16989084621203934, "grad_norm": 0.3543424606323242, "learning_rate": 8e-05, "loss": 1.715, "step": 1572 }, { "epoch": 0.16999891926942612, "grad_norm": 0.3879331648349762, "learning_rate": 8e-05, "loss": 1.9224, "step": 1573 }, { "epoch": 0.17010699232681292, "grad_norm": 0.39529386162757874, "learning_rate": 8e-05, "loss": 1.7981, "step": 1574 }, { "epoch": 0.17021506538419973, "grad_norm": 0.3868407905101776, "learning_rate": 8e-05, "loss": 1.8664, "step": 1575 }, { "epoch": 0.1703231384415865, "grad_norm": 0.3503515124320984, "learning_rate": 8e-05, "loss": 1.7718, "step": 1576 }, { "epoch": 0.1704312114989733, "grad_norm": 0.33907774090766907, "learning_rate": 8e-05, "loss": 1.7033, "step": 1577 }, { "epoch": 0.1705392845563601, "grad_norm": 0.34026020765304565, "learning_rate": 8e-05, "loss": 1.6399, "step": 1578 }, { "epoch": 0.1706473576137469, "grad_norm": 0.4544883072376251, "learning_rate": 8e-05, "loss": 1.6796, "step": 1579 }, { "epoch": 0.1707554306711337, "grad_norm": 0.3494049608707428, "learning_rate": 8e-05, "loss": 1.7372, "step": 1580 }, { "epoch": 0.17086350372852047, "grad_norm": 0.3605392277240753, "learning_rate": 8e-05, "loss": 1.7685, "step": 1581 }, { "epoch": 0.17097157678590727, "grad_norm": 0.3782109320163727, "learning_rate": 8e-05, "loss": 1.7761, "step": 1582 }, { "epoch": 0.17107964984329407, "grad_norm": 0.39154326915740967, "learning_rate": 8e-05, "loss": 1.7806, "step": 1583 }, { "epoch": 0.17118772290068085, "grad_norm": 0.36011967062950134, "learning_rate": 8e-05, "loss": 1.6658, "step": 1584 }, { "epoch": 0.17129579595806765, "grad_norm": 0.3702276051044464, "learning_rate": 8e-05, "loss": 1.5475, "step": 1585 }, { "epoch": 0.17140386901545446, "grad_norm": 0.3594640791416168, "learning_rate": 8e-05, "loss": 1.7196, "step": 1586 }, { "epoch": 0.17151194207284123, "grad_norm": 0.36223164200782776, "learning_rate": 8e-05, "loss": 1.8273, "step": 1587 }, { "epoch": 0.17162001513022804, "grad_norm": 0.34349924325942993, "learning_rate": 8e-05, "loss": 1.7268, "step": 1588 }, { "epoch": 0.17172808818761484, "grad_norm": 0.3908827602863312, "learning_rate": 8e-05, "loss": 1.8184, "step": 1589 }, { "epoch": 0.17183616124500162, "grad_norm": 0.35485586524009705, "learning_rate": 8e-05, "loss": 1.6687, "step": 1590 }, { "epoch": 0.17194423430238842, "grad_norm": 0.37867337465286255, "learning_rate": 8e-05, "loss": 1.7827, "step": 1591 }, { "epoch": 0.1720523073597752, "grad_norm": 0.38368913531303406, "learning_rate": 8e-05, "loss": 1.8279, "step": 1592 }, { "epoch": 0.172160380417162, "grad_norm": 0.3789379894733429, "learning_rate": 8e-05, "loss": 1.6387, "step": 1593 }, { "epoch": 0.1722684534745488, "grad_norm": 0.3658318817615509, "learning_rate": 8e-05, "loss": 1.7845, "step": 1594 }, { "epoch": 0.17237652653193558, "grad_norm": 0.36771678924560547, "learning_rate": 8e-05, "loss": 1.7356, "step": 1595 }, { "epoch": 0.17248459958932238, "grad_norm": 0.36121058464050293, "learning_rate": 8e-05, "loss": 1.7365, "step": 1596 }, { "epoch": 0.1725926726467092, "grad_norm": 0.36337608098983765, "learning_rate": 8e-05, "loss": 1.8057, "step": 1597 }, { "epoch": 0.17270074570409596, "grad_norm": 0.36104637384414673, "learning_rate": 8e-05, "loss": 1.7135, "step": 1598 }, { "epoch": 0.17280881876148277, "grad_norm": 0.369521826505661, "learning_rate": 8e-05, "loss": 1.8027, "step": 1599 }, { "epoch": 0.17291689181886954, "grad_norm": 0.3675416111946106, "learning_rate": 8e-05, "loss": 1.6951, "step": 1600 }, { "epoch": 0.17302496487625635, "grad_norm": 0.3630196154117584, "learning_rate": 8e-05, "loss": 1.9198, "step": 1601 }, { "epoch": 0.17313303793364315, "grad_norm": 0.3838110864162445, "learning_rate": 8e-05, "loss": 1.9784, "step": 1602 }, { "epoch": 0.17324111099102993, "grad_norm": 0.34828025102615356, "learning_rate": 8e-05, "loss": 1.6791, "step": 1603 }, { "epoch": 0.17334918404841673, "grad_norm": 0.376072496175766, "learning_rate": 8e-05, "loss": 1.8649, "step": 1604 }, { "epoch": 0.17345725710580354, "grad_norm": 0.38232260942459106, "learning_rate": 8e-05, "loss": 1.9361, "step": 1605 }, { "epoch": 0.1735653301631903, "grad_norm": 0.360359251499176, "learning_rate": 8e-05, "loss": 1.6816, "step": 1606 }, { "epoch": 0.17367340322057712, "grad_norm": 0.34956425428390503, "learning_rate": 8e-05, "loss": 1.6475, "step": 1607 }, { "epoch": 0.1737814762779639, "grad_norm": 0.40701553225517273, "learning_rate": 8e-05, "loss": 1.888, "step": 1608 }, { "epoch": 0.1738895493353507, "grad_norm": 0.37468236684799194, "learning_rate": 8e-05, "loss": 1.8456, "step": 1609 }, { "epoch": 0.1739976223927375, "grad_norm": 0.3446200489997864, "learning_rate": 8e-05, "loss": 1.7896, "step": 1610 }, { "epoch": 0.17410569545012428, "grad_norm": 0.37653806805610657, "learning_rate": 8e-05, "loss": 1.8452, "step": 1611 }, { "epoch": 0.17421376850751108, "grad_norm": 0.35356035828590393, "learning_rate": 8e-05, "loss": 1.7478, "step": 1612 }, { "epoch": 0.17432184156489788, "grad_norm": 0.3565081059932709, "learning_rate": 8e-05, "loss": 1.8095, "step": 1613 }, { "epoch": 0.17442991462228466, "grad_norm": 0.3607056140899658, "learning_rate": 8e-05, "loss": 1.6408, "step": 1614 }, { "epoch": 0.17453798767967146, "grad_norm": 0.3345412611961365, "learning_rate": 8e-05, "loss": 1.7174, "step": 1615 }, { "epoch": 0.17464606073705824, "grad_norm": 0.3666120767593384, "learning_rate": 8e-05, "loss": 1.8134, "step": 1616 }, { "epoch": 0.17475413379444504, "grad_norm": 0.34048354625701904, "learning_rate": 8e-05, "loss": 1.7696, "step": 1617 }, { "epoch": 0.17486220685183185, "grad_norm": 0.35646718740463257, "learning_rate": 8e-05, "loss": 1.9065, "step": 1618 }, { "epoch": 0.17497027990921862, "grad_norm": 0.42236730456352234, "learning_rate": 8e-05, "loss": 1.9125, "step": 1619 }, { "epoch": 0.17507835296660543, "grad_norm": 0.40540653467178345, "learning_rate": 8e-05, "loss": 1.8505, "step": 1620 }, { "epoch": 0.17518642602399223, "grad_norm": 0.331855446100235, "learning_rate": 8e-05, "loss": 1.7593, "step": 1621 }, { "epoch": 0.175294499081379, "grad_norm": 0.36744189262390137, "learning_rate": 8e-05, "loss": 1.8875, "step": 1622 }, { "epoch": 0.1754025721387658, "grad_norm": 0.3942550718784332, "learning_rate": 8e-05, "loss": 1.956, "step": 1623 }, { "epoch": 0.1755106451961526, "grad_norm": 0.3567750155925751, "learning_rate": 8e-05, "loss": 1.7495, "step": 1624 }, { "epoch": 0.1756187182535394, "grad_norm": 0.36972054839134216, "learning_rate": 8e-05, "loss": 1.7481, "step": 1625 }, { "epoch": 0.1757267913109262, "grad_norm": 0.3746875822544098, "learning_rate": 8e-05, "loss": 1.8802, "step": 1626 }, { "epoch": 0.17583486436831297, "grad_norm": 0.3661128878593445, "learning_rate": 8e-05, "loss": 1.7955, "step": 1627 }, { "epoch": 0.17594293742569977, "grad_norm": 0.35059115290641785, "learning_rate": 8e-05, "loss": 1.8725, "step": 1628 }, { "epoch": 0.17605101048308658, "grad_norm": 0.36802592873573303, "learning_rate": 8e-05, "loss": 1.661, "step": 1629 }, { "epoch": 0.17615908354047335, "grad_norm": 0.36493128538131714, "learning_rate": 8e-05, "loss": 1.8347, "step": 1630 }, { "epoch": 0.17626715659786016, "grad_norm": 0.3848332464694977, "learning_rate": 8e-05, "loss": 1.7785, "step": 1631 }, { "epoch": 0.17637522965524693, "grad_norm": 0.3916698098182678, "learning_rate": 8e-05, "loss": 1.7927, "step": 1632 }, { "epoch": 0.17648330271263374, "grad_norm": 0.45152509212493896, "learning_rate": 8e-05, "loss": 2.045, "step": 1633 }, { "epoch": 0.17659137577002054, "grad_norm": 0.4029836058616638, "learning_rate": 8e-05, "loss": 1.9389, "step": 1634 }, { "epoch": 0.17669944882740732, "grad_norm": 0.4467923939228058, "learning_rate": 8e-05, "loss": 2.0114, "step": 1635 }, { "epoch": 0.17680752188479412, "grad_norm": 0.4108101725578308, "learning_rate": 8e-05, "loss": 1.7711, "step": 1636 }, { "epoch": 0.17691559494218093, "grad_norm": 0.38180819153785706, "learning_rate": 8e-05, "loss": 1.7848, "step": 1637 }, { "epoch": 0.1770236679995677, "grad_norm": 0.3952118754386902, "learning_rate": 8e-05, "loss": 1.8248, "step": 1638 }, { "epoch": 0.1771317410569545, "grad_norm": 0.38845109939575195, "learning_rate": 8e-05, "loss": 1.7078, "step": 1639 }, { "epoch": 0.17723981411434128, "grad_norm": 0.359057754278183, "learning_rate": 8e-05, "loss": 1.781, "step": 1640 }, { "epoch": 0.17734788717172809, "grad_norm": 0.34535807371139526, "learning_rate": 8e-05, "loss": 1.7322, "step": 1641 }, { "epoch": 0.1774559602291149, "grad_norm": 0.37608200311660767, "learning_rate": 8e-05, "loss": 1.7808, "step": 1642 }, { "epoch": 0.17756403328650167, "grad_norm": 0.4053289592266083, "learning_rate": 8e-05, "loss": 1.9428, "step": 1643 }, { "epoch": 0.17767210634388847, "grad_norm": 0.3741133511066437, "learning_rate": 8e-05, "loss": 1.8434, "step": 1644 }, { "epoch": 0.17778017940127527, "grad_norm": 0.34029701352119446, "learning_rate": 8e-05, "loss": 1.7948, "step": 1645 }, { "epoch": 0.17788825245866205, "grad_norm": 0.35523542761802673, "learning_rate": 8e-05, "loss": 1.7983, "step": 1646 }, { "epoch": 0.17799632551604885, "grad_norm": 0.35122838616371155, "learning_rate": 8e-05, "loss": 1.6771, "step": 1647 }, { "epoch": 0.17810439857343563, "grad_norm": 0.36682063341140747, "learning_rate": 8e-05, "loss": 1.7303, "step": 1648 }, { "epoch": 0.17821247163082243, "grad_norm": 0.34502407908439636, "learning_rate": 8e-05, "loss": 1.6997, "step": 1649 }, { "epoch": 0.17832054468820924, "grad_norm": 0.36331263184547424, "learning_rate": 8e-05, "loss": 1.7576, "step": 1650 }, { "epoch": 0.178428617745596, "grad_norm": 0.3590247929096222, "learning_rate": 8e-05, "loss": 1.6528, "step": 1651 }, { "epoch": 0.17853669080298282, "grad_norm": 0.3526703119277954, "learning_rate": 8e-05, "loss": 1.706, "step": 1652 }, { "epoch": 0.17864476386036962, "grad_norm": 0.36700600385665894, "learning_rate": 8e-05, "loss": 1.8348, "step": 1653 }, { "epoch": 0.1787528369177564, "grad_norm": 0.37900376319885254, "learning_rate": 8e-05, "loss": 1.7998, "step": 1654 }, { "epoch": 0.1788609099751432, "grad_norm": 0.3786175847053528, "learning_rate": 8e-05, "loss": 1.7667, "step": 1655 }, { "epoch": 0.17896898303252998, "grad_norm": 0.3788127303123474, "learning_rate": 8e-05, "loss": 1.7213, "step": 1656 }, { "epoch": 0.17907705608991678, "grad_norm": 0.3550642728805542, "learning_rate": 8e-05, "loss": 1.6294, "step": 1657 }, { "epoch": 0.17918512914730358, "grad_norm": 0.372003972530365, "learning_rate": 8e-05, "loss": 1.7615, "step": 1658 }, { "epoch": 0.17929320220469036, "grad_norm": 0.35628843307495117, "learning_rate": 8e-05, "loss": 1.7282, "step": 1659 }, { "epoch": 0.17940127526207716, "grad_norm": 0.3718128204345703, "learning_rate": 8e-05, "loss": 1.6831, "step": 1660 }, { "epoch": 0.17950934831946397, "grad_norm": 0.3651917278766632, "learning_rate": 8e-05, "loss": 1.8679, "step": 1661 }, { "epoch": 0.17961742137685074, "grad_norm": 0.40477707982063293, "learning_rate": 8e-05, "loss": 1.9552, "step": 1662 }, { "epoch": 0.17972549443423755, "grad_norm": 0.33563488721847534, "learning_rate": 8e-05, "loss": 1.7413, "step": 1663 }, { "epoch": 0.17983356749162432, "grad_norm": 0.35286417603492737, "learning_rate": 8e-05, "loss": 1.649, "step": 1664 }, { "epoch": 0.17994164054901113, "grad_norm": 0.39628294110298157, "learning_rate": 8e-05, "loss": 1.6988, "step": 1665 }, { "epoch": 0.18004971360639793, "grad_norm": 0.3455020487308502, "learning_rate": 8e-05, "loss": 1.7121, "step": 1666 }, { "epoch": 0.1801577866637847, "grad_norm": 0.39466869831085205, "learning_rate": 8e-05, "loss": 1.6827, "step": 1667 }, { "epoch": 0.1802658597211715, "grad_norm": 0.3614978790283203, "learning_rate": 8e-05, "loss": 1.814, "step": 1668 }, { "epoch": 0.18037393277855832, "grad_norm": 0.36236196756362915, "learning_rate": 8e-05, "loss": 1.7807, "step": 1669 }, { "epoch": 0.1804820058359451, "grad_norm": 0.3276165723800659, "learning_rate": 8e-05, "loss": 1.5727, "step": 1670 }, { "epoch": 0.1805900788933319, "grad_norm": 0.3774045407772064, "learning_rate": 8e-05, "loss": 1.7476, "step": 1671 }, { "epoch": 0.1806981519507187, "grad_norm": 0.34874510765075684, "learning_rate": 8e-05, "loss": 1.6803, "step": 1672 }, { "epoch": 0.18080622500810548, "grad_norm": 0.36427995562553406, "learning_rate": 8e-05, "loss": 1.7486, "step": 1673 }, { "epoch": 0.18091429806549228, "grad_norm": 0.38381022214889526, "learning_rate": 8e-05, "loss": 1.7939, "step": 1674 }, { "epoch": 0.18102237112287906, "grad_norm": 0.3950136601924896, "learning_rate": 8e-05, "loss": 1.9148, "step": 1675 }, { "epoch": 0.18113044418026586, "grad_norm": 0.37090596556663513, "learning_rate": 8e-05, "loss": 1.8078, "step": 1676 }, { "epoch": 0.18123851723765266, "grad_norm": 0.36812740564346313, "learning_rate": 8e-05, "loss": 1.7293, "step": 1677 }, { "epoch": 0.18134659029503944, "grad_norm": 0.3782331645488739, "learning_rate": 8e-05, "loss": 1.8253, "step": 1678 }, { "epoch": 0.18145466335242624, "grad_norm": 0.3781827986240387, "learning_rate": 8e-05, "loss": 1.9345, "step": 1679 }, { "epoch": 0.18156273640981305, "grad_norm": 0.37168988585472107, "learning_rate": 8e-05, "loss": 1.7043, "step": 1680 }, { "epoch": 0.18167080946719982, "grad_norm": 0.34382009506225586, "learning_rate": 8e-05, "loss": 1.5828, "step": 1681 }, { "epoch": 0.18177888252458663, "grad_norm": 0.3758091628551483, "learning_rate": 8e-05, "loss": 1.8229, "step": 1682 }, { "epoch": 0.1818869555819734, "grad_norm": 0.3712696135044098, "learning_rate": 8e-05, "loss": 1.8316, "step": 1683 }, { "epoch": 0.1819950286393602, "grad_norm": 0.36939799785614014, "learning_rate": 8e-05, "loss": 1.8414, "step": 1684 }, { "epoch": 0.182103101696747, "grad_norm": 0.36691927909851074, "learning_rate": 8e-05, "loss": 1.7739, "step": 1685 }, { "epoch": 0.1822111747541338, "grad_norm": 0.3789370059967041, "learning_rate": 8e-05, "loss": 1.8676, "step": 1686 }, { "epoch": 0.1823192478115206, "grad_norm": 0.3779565393924713, "learning_rate": 8e-05, "loss": 1.6768, "step": 1687 }, { "epoch": 0.1824273208689074, "grad_norm": 0.34127485752105713, "learning_rate": 8e-05, "loss": 1.7272, "step": 1688 }, { "epoch": 0.18253539392629417, "grad_norm": 0.3747107684612274, "learning_rate": 8e-05, "loss": 1.6989, "step": 1689 }, { "epoch": 0.18264346698368097, "grad_norm": 0.35757023096084595, "learning_rate": 8e-05, "loss": 1.7568, "step": 1690 }, { "epoch": 0.18275154004106775, "grad_norm": 0.3674648404121399, "learning_rate": 8e-05, "loss": 1.8291, "step": 1691 }, { "epoch": 0.18285961309845455, "grad_norm": 0.38322946429252625, "learning_rate": 8e-05, "loss": 1.9127, "step": 1692 }, { "epoch": 0.18296768615584136, "grad_norm": 0.33984994888305664, "learning_rate": 8e-05, "loss": 1.7354, "step": 1693 }, { "epoch": 0.18307575921322813, "grad_norm": 0.3492932617664337, "learning_rate": 8e-05, "loss": 1.817, "step": 1694 }, { "epoch": 0.18318383227061494, "grad_norm": 0.3868374526500702, "learning_rate": 8e-05, "loss": 1.9592, "step": 1695 }, { "epoch": 0.18329190532800174, "grad_norm": 0.3798457682132721, "learning_rate": 8e-05, "loss": 1.9557, "step": 1696 }, { "epoch": 0.18339997838538852, "grad_norm": 0.341256707906723, "learning_rate": 8e-05, "loss": 1.5615, "step": 1697 }, { "epoch": 0.18350805144277532, "grad_norm": 0.36626920104026794, "learning_rate": 8e-05, "loss": 1.557, "step": 1698 }, { "epoch": 0.1836161245001621, "grad_norm": 0.37801268696784973, "learning_rate": 8e-05, "loss": 1.9493, "step": 1699 }, { "epoch": 0.1837241975575489, "grad_norm": 0.39826786518096924, "learning_rate": 8e-05, "loss": 1.8467, "step": 1700 }, { "epoch": 0.1838322706149357, "grad_norm": 0.35098350048065186, "learning_rate": 8e-05, "loss": 1.7381, "step": 1701 }, { "epoch": 0.18394034367232248, "grad_norm": 0.36847978830337524, "learning_rate": 8e-05, "loss": 1.8027, "step": 1702 }, { "epoch": 0.18404841672970929, "grad_norm": 0.3955918252468109, "learning_rate": 8e-05, "loss": 1.7525, "step": 1703 }, { "epoch": 0.1841564897870961, "grad_norm": 0.3944500982761383, "learning_rate": 8e-05, "loss": 1.8408, "step": 1704 }, { "epoch": 0.18426456284448287, "grad_norm": 0.34581539034843445, "learning_rate": 8e-05, "loss": 1.5763, "step": 1705 }, { "epoch": 0.18437263590186967, "grad_norm": 0.37280723452568054, "learning_rate": 8e-05, "loss": 1.7851, "step": 1706 }, { "epoch": 0.18448070895925645, "grad_norm": 0.32942456007003784, "learning_rate": 8e-05, "loss": 1.6265, "step": 1707 }, { "epoch": 0.18458878201664325, "grad_norm": 0.35244083404541016, "learning_rate": 8e-05, "loss": 1.7128, "step": 1708 }, { "epoch": 0.18469685507403005, "grad_norm": 0.36334046721458435, "learning_rate": 8e-05, "loss": 1.8308, "step": 1709 }, { "epoch": 0.18480492813141683, "grad_norm": 0.3682284653186798, "learning_rate": 8e-05, "loss": 1.7161, "step": 1710 }, { "epoch": 0.18491300118880363, "grad_norm": 0.3617183268070221, "learning_rate": 8e-05, "loss": 1.7895, "step": 1711 }, { "epoch": 0.18502107424619044, "grad_norm": 0.3655518591403961, "learning_rate": 8e-05, "loss": 1.7255, "step": 1712 }, { "epoch": 0.1851291473035772, "grad_norm": 0.3500616252422333, "learning_rate": 8e-05, "loss": 1.769, "step": 1713 }, { "epoch": 0.18523722036096402, "grad_norm": 0.3690573275089264, "learning_rate": 8e-05, "loss": 1.6323, "step": 1714 }, { "epoch": 0.1853452934183508, "grad_norm": 0.3751242458820343, "learning_rate": 8e-05, "loss": 1.801, "step": 1715 }, { "epoch": 0.1854533664757376, "grad_norm": 0.3810383081436157, "learning_rate": 8e-05, "loss": 1.7926, "step": 1716 }, { "epoch": 0.1855614395331244, "grad_norm": 0.36056819558143616, "learning_rate": 8e-05, "loss": 1.6279, "step": 1717 }, { "epoch": 0.18566951259051118, "grad_norm": 0.3514404594898224, "learning_rate": 8e-05, "loss": 1.8355, "step": 1718 }, { "epoch": 0.18577758564789798, "grad_norm": 0.36354658007621765, "learning_rate": 8e-05, "loss": 1.7155, "step": 1719 }, { "epoch": 0.18588565870528478, "grad_norm": 0.3387967050075531, "learning_rate": 8e-05, "loss": 1.7008, "step": 1720 }, { "epoch": 0.18599373176267156, "grad_norm": 0.35997846722602844, "learning_rate": 8e-05, "loss": 1.7762, "step": 1721 }, { "epoch": 0.18610180482005836, "grad_norm": 0.3943254351615906, "learning_rate": 8e-05, "loss": 1.8421, "step": 1722 }, { "epoch": 0.18620987787744514, "grad_norm": 0.3703499734401703, "learning_rate": 8e-05, "loss": 1.7988, "step": 1723 }, { "epoch": 0.18631795093483194, "grad_norm": 0.3872607946395874, "learning_rate": 8e-05, "loss": 1.8857, "step": 1724 }, { "epoch": 0.18642602399221875, "grad_norm": 0.3822230398654938, "learning_rate": 8e-05, "loss": 1.8202, "step": 1725 }, { "epoch": 0.18653409704960552, "grad_norm": 0.43150418996810913, "learning_rate": 8e-05, "loss": 1.9394, "step": 1726 }, { "epoch": 0.18664217010699233, "grad_norm": 0.36196598410606384, "learning_rate": 8e-05, "loss": 1.738, "step": 1727 }, { "epoch": 0.18675024316437913, "grad_norm": 0.4014938473701477, "learning_rate": 8e-05, "loss": 1.8033, "step": 1728 }, { "epoch": 0.1868583162217659, "grad_norm": 0.3539465665817261, "learning_rate": 8e-05, "loss": 1.7229, "step": 1729 }, { "epoch": 0.1869663892791527, "grad_norm": 0.3777218461036682, "learning_rate": 8e-05, "loss": 1.7343, "step": 1730 }, { "epoch": 0.1870744623365395, "grad_norm": 0.4082215130329132, "learning_rate": 8e-05, "loss": 1.6875, "step": 1731 }, { "epoch": 0.1871825353939263, "grad_norm": 0.37269142270088196, "learning_rate": 8e-05, "loss": 1.8597, "step": 1732 }, { "epoch": 0.1872906084513131, "grad_norm": 0.3698691427707672, "learning_rate": 8e-05, "loss": 1.7293, "step": 1733 }, { "epoch": 0.18739868150869987, "grad_norm": 0.35455918312072754, "learning_rate": 8e-05, "loss": 1.8025, "step": 1734 }, { "epoch": 0.18750675456608668, "grad_norm": 0.3900121748447418, "learning_rate": 8e-05, "loss": 1.8418, "step": 1735 }, { "epoch": 0.18761482762347348, "grad_norm": 0.3949868083000183, "learning_rate": 8e-05, "loss": 1.9188, "step": 1736 }, { "epoch": 0.18772290068086026, "grad_norm": 0.3427894115447998, "learning_rate": 8e-05, "loss": 1.7221, "step": 1737 }, { "epoch": 0.18783097373824706, "grad_norm": 0.3467901349067688, "learning_rate": 8e-05, "loss": 1.7529, "step": 1738 }, { "epoch": 0.18793904679563384, "grad_norm": 0.3490123748779297, "learning_rate": 8e-05, "loss": 1.792, "step": 1739 }, { "epoch": 0.18804711985302064, "grad_norm": 0.4113178551197052, "learning_rate": 8e-05, "loss": 1.8467, "step": 1740 }, { "epoch": 0.18815519291040744, "grad_norm": 0.3516266345977783, "learning_rate": 8e-05, "loss": 1.6673, "step": 1741 }, { "epoch": 0.18826326596779422, "grad_norm": 0.35637328028678894, "learning_rate": 8e-05, "loss": 1.717, "step": 1742 }, { "epoch": 0.18837133902518102, "grad_norm": 0.3567594289779663, "learning_rate": 8e-05, "loss": 1.714, "step": 1743 }, { "epoch": 0.18847941208256783, "grad_norm": 0.3844467103481293, "learning_rate": 8e-05, "loss": 1.8178, "step": 1744 }, { "epoch": 0.1885874851399546, "grad_norm": 0.37149736285209656, "learning_rate": 8e-05, "loss": 1.8056, "step": 1745 }, { "epoch": 0.1886955581973414, "grad_norm": 0.36744555830955505, "learning_rate": 8e-05, "loss": 1.6588, "step": 1746 }, { "epoch": 0.18880363125472818, "grad_norm": 0.3590886890888214, "learning_rate": 8e-05, "loss": 1.8339, "step": 1747 }, { "epoch": 0.188911704312115, "grad_norm": 0.38177254796028137, "learning_rate": 8e-05, "loss": 1.7627, "step": 1748 }, { "epoch": 0.1890197773695018, "grad_norm": 0.3755657970905304, "learning_rate": 8e-05, "loss": 1.809, "step": 1749 }, { "epoch": 0.18912785042688857, "grad_norm": 0.34013307094573975, "learning_rate": 8e-05, "loss": 1.7792, "step": 1750 }, { "epoch": 0.18923592348427537, "grad_norm": 0.3533726632595062, "learning_rate": 8e-05, "loss": 1.6965, "step": 1751 }, { "epoch": 0.18934399654166217, "grad_norm": 0.3901097774505615, "learning_rate": 8e-05, "loss": 1.8536, "step": 1752 }, { "epoch": 0.18945206959904895, "grad_norm": 0.355742871761322, "learning_rate": 8e-05, "loss": 1.6134, "step": 1753 }, { "epoch": 0.18956014265643575, "grad_norm": 0.3505704998970032, "learning_rate": 8e-05, "loss": 1.7236, "step": 1754 }, { "epoch": 0.18966821571382256, "grad_norm": 0.3514764904975891, "learning_rate": 8e-05, "loss": 1.7449, "step": 1755 }, { "epoch": 0.18977628877120933, "grad_norm": 0.33720675110816956, "learning_rate": 8e-05, "loss": 1.733, "step": 1756 }, { "epoch": 0.18988436182859614, "grad_norm": 0.4030938148498535, "learning_rate": 8e-05, "loss": 1.8353, "step": 1757 }, { "epoch": 0.1899924348859829, "grad_norm": 0.3569396734237671, "learning_rate": 8e-05, "loss": 1.7472, "step": 1758 }, { "epoch": 0.19010050794336972, "grad_norm": 0.374889999628067, "learning_rate": 8e-05, "loss": 1.8444, "step": 1759 }, { "epoch": 0.19020858100075652, "grad_norm": 0.3418726921081543, "learning_rate": 8e-05, "loss": 1.5447, "step": 1760 }, { "epoch": 0.1903166540581433, "grad_norm": 0.3392559289932251, "learning_rate": 8e-05, "loss": 1.4894, "step": 1761 }, { "epoch": 0.1904247271155301, "grad_norm": 0.3604184091091156, "learning_rate": 8e-05, "loss": 1.8812, "step": 1762 }, { "epoch": 0.1905328001729169, "grad_norm": 0.35275551676750183, "learning_rate": 8e-05, "loss": 1.7426, "step": 1763 }, { "epoch": 0.19064087323030368, "grad_norm": 0.3618411123752594, "learning_rate": 8e-05, "loss": 1.6833, "step": 1764 }, { "epoch": 0.19074894628769049, "grad_norm": 0.3477647602558136, "learning_rate": 8e-05, "loss": 1.8296, "step": 1765 }, { "epoch": 0.19085701934507726, "grad_norm": 0.3734968304634094, "learning_rate": 8e-05, "loss": 1.8205, "step": 1766 }, { "epoch": 0.19096509240246407, "grad_norm": 0.3685232996940613, "learning_rate": 8e-05, "loss": 1.736, "step": 1767 }, { "epoch": 0.19107316545985087, "grad_norm": 0.4022296667098999, "learning_rate": 8e-05, "loss": 1.8617, "step": 1768 }, { "epoch": 0.19118123851723764, "grad_norm": 0.3684019446372986, "learning_rate": 8e-05, "loss": 1.865, "step": 1769 }, { "epoch": 0.19128931157462445, "grad_norm": 0.3642842471599579, "learning_rate": 8e-05, "loss": 1.8401, "step": 1770 }, { "epoch": 0.19139738463201125, "grad_norm": 0.35271260142326355, "learning_rate": 8e-05, "loss": 1.6759, "step": 1771 }, { "epoch": 0.19150545768939803, "grad_norm": 0.36002832651138306, "learning_rate": 8e-05, "loss": 1.7049, "step": 1772 }, { "epoch": 0.19161353074678483, "grad_norm": 0.33967751264572144, "learning_rate": 8e-05, "loss": 1.7104, "step": 1773 }, { "epoch": 0.1917216038041716, "grad_norm": 0.35905686020851135, "learning_rate": 8e-05, "loss": 1.6578, "step": 1774 }, { "epoch": 0.1918296768615584, "grad_norm": 0.3566742539405823, "learning_rate": 8e-05, "loss": 1.7877, "step": 1775 }, { "epoch": 0.19193774991894522, "grad_norm": 0.35294750332832336, "learning_rate": 8e-05, "loss": 1.8371, "step": 1776 }, { "epoch": 0.192045822976332, "grad_norm": 0.3792085647583008, "learning_rate": 8e-05, "loss": 1.8229, "step": 1777 }, { "epoch": 0.1921538960337188, "grad_norm": 0.36063823103904724, "learning_rate": 8e-05, "loss": 1.7812, "step": 1778 }, { "epoch": 0.1922619690911056, "grad_norm": 0.3558347225189209, "learning_rate": 8e-05, "loss": 1.6367, "step": 1779 }, { "epoch": 0.19237004214849238, "grad_norm": 0.39099588990211487, "learning_rate": 8e-05, "loss": 1.826, "step": 1780 }, { "epoch": 0.19247811520587918, "grad_norm": 0.35245659947395325, "learning_rate": 8e-05, "loss": 1.6881, "step": 1781 }, { "epoch": 0.19258618826326596, "grad_norm": 0.38026079535484314, "learning_rate": 8e-05, "loss": 1.8844, "step": 1782 }, { "epoch": 0.19269426132065276, "grad_norm": 0.35403531789779663, "learning_rate": 8e-05, "loss": 1.661, "step": 1783 }, { "epoch": 0.19280233437803956, "grad_norm": 0.3622284531593323, "learning_rate": 8e-05, "loss": 1.6176, "step": 1784 }, { "epoch": 0.19291040743542634, "grad_norm": 0.3586704432964325, "learning_rate": 8e-05, "loss": 1.6468, "step": 1785 }, { "epoch": 0.19301848049281314, "grad_norm": 0.3440355360507965, "learning_rate": 8e-05, "loss": 1.5569, "step": 1786 }, { "epoch": 0.19312655355019995, "grad_norm": 0.3519052565097809, "learning_rate": 8e-05, "loss": 1.6826, "step": 1787 }, { "epoch": 0.19323462660758672, "grad_norm": 0.4002799689769745, "learning_rate": 8e-05, "loss": 1.9489, "step": 1788 }, { "epoch": 0.19334269966497353, "grad_norm": 0.37021762132644653, "learning_rate": 8e-05, "loss": 1.8333, "step": 1789 }, { "epoch": 0.1934507727223603, "grad_norm": 0.37534624338150024, "learning_rate": 8e-05, "loss": 1.8305, "step": 1790 }, { "epoch": 0.1935588457797471, "grad_norm": 0.35690805315971375, "learning_rate": 8e-05, "loss": 1.5306, "step": 1791 }, { "epoch": 0.1936669188371339, "grad_norm": 0.3545568585395813, "learning_rate": 8e-05, "loss": 1.6562, "step": 1792 }, { "epoch": 0.1937749918945207, "grad_norm": 0.3569551706314087, "learning_rate": 8e-05, "loss": 1.7533, "step": 1793 }, { "epoch": 0.1938830649519075, "grad_norm": 0.4669347405433655, "learning_rate": 8e-05, "loss": 1.8666, "step": 1794 }, { "epoch": 0.1939911380092943, "grad_norm": 0.3874759078025818, "learning_rate": 8e-05, "loss": 1.7271, "step": 1795 }, { "epoch": 0.19409921106668107, "grad_norm": 0.3388046622276306, "learning_rate": 8e-05, "loss": 1.6284, "step": 1796 }, { "epoch": 0.19420728412406787, "grad_norm": 0.3450491726398468, "learning_rate": 8e-05, "loss": 1.6565, "step": 1797 }, { "epoch": 0.19431535718145465, "grad_norm": 0.36635780334472656, "learning_rate": 8e-05, "loss": 1.8437, "step": 1798 }, { "epoch": 0.19442343023884145, "grad_norm": 0.3815009295940399, "learning_rate": 8e-05, "loss": 1.7591, "step": 1799 }, { "epoch": 0.19453150329622826, "grad_norm": 0.3911726176738739, "learning_rate": 8e-05, "loss": 1.9081, "step": 1800 }, { "epoch": 0.19463957635361503, "grad_norm": 0.39948198199272156, "learning_rate": 8e-05, "loss": 1.8739, "step": 1801 }, { "epoch": 0.19474764941100184, "grad_norm": 0.3867413401603699, "learning_rate": 8e-05, "loss": 1.716, "step": 1802 }, { "epoch": 0.19485572246838864, "grad_norm": 0.4122034013271332, "learning_rate": 8e-05, "loss": 1.9253, "step": 1803 }, { "epoch": 0.19496379552577542, "grad_norm": 0.34969455003738403, "learning_rate": 8e-05, "loss": 1.5936, "step": 1804 }, { "epoch": 0.19507186858316222, "grad_norm": 0.35876497626304626, "learning_rate": 8e-05, "loss": 1.6905, "step": 1805 }, { "epoch": 0.195179941640549, "grad_norm": 0.33905214071273804, "learning_rate": 8e-05, "loss": 1.6594, "step": 1806 }, { "epoch": 0.1952880146979358, "grad_norm": 0.35875120759010315, "learning_rate": 8e-05, "loss": 1.7224, "step": 1807 }, { "epoch": 0.1953960877553226, "grad_norm": 0.3817675709724426, "learning_rate": 8e-05, "loss": 1.8758, "step": 1808 }, { "epoch": 0.19550416081270938, "grad_norm": 0.37900206446647644, "learning_rate": 8e-05, "loss": 1.8086, "step": 1809 }, { "epoch": 0.19561223387009619, "grad_norm": 0.38451164960861206, "learning_rate": 8e-05, "loss": 1.9758, "step": 1810 }, { "epoch": 0.195720306927483, "grad_norm": 0.36660683155059814, "learning_rate": 8e-05, "loss": 1.9801, "step": 1811 }, { "epoch": 0.19582837998486977, "grad_norm": 0.37666767835617065, "learning_rate": 8e-05, "loss": 1.6089, "step": 1812 }, { "epoch": 0.19593645304225657, "grad_norm": 0.3671923279762268, "learning_rate": 8e-05, "loss": 1.4948, "step": 1813 }, { "epoch": 0.19604452609964335, "grad_norm": 0.3429158627986908, "learning_rate": 8e-05, "loss": 1.6708, "step": 1814 }, { "epoch": 0.19615259915703015, "grad_norm": 0.3746236562728882, "learning_rate": 8e-05, "loss": 1.8111, "step": 1815 }, { "epoch": 0.19626067221441695, "grad_norm": 0.3776658773422241, "learning_rate": 8e-05, "loss": 1.6249, "step": 1816 }, { "epoch": 0.19636874527180373, "grad_norm": 0.4022580087184906, "learning_rate": 8e-05, "loss": 1.7203, "step": 1817 }, { "epoch": 0.19647681832919053, "grad_norm": 0.40486854314804077, "learning_rate": 8e-05, "loss": 1.9984, "step": 1818 }, { "epoch": 0.19658489138657734, "grad_norm": 0.3534870445728302, "learning_rate": 8e-05, "loss": 1.6254, "step": 1819 }, { "epoch": 0.1966929644439641, "grad_norm": 0.35138943791389465, "learning_rate": 8e-05, "loss": 1.6122, "step": 1820 }, { "epoch": 0.19680103750135092, "grad_norm": 0.36506959795951843, "learning_rate": 8e-05, "loss": 1.8256, "step": 1821 }, { "epoch": 0.1969091105587377, "grad_norm": 0.40358293056488037, "learning_rate": 8e-05, "loss": 2.0094, "step": 1822 }, { "epoch": 0.1970171836161245, "grad_norm": 0.3913581073284149, "learning_rate": 8e-05, "loss": 1.6099, "step": 1823 }, { "epoch": 0.1971252566735113, "grad_norm": 0.3925180435180664, "learning_rate": 8e-05, "loss": 1.9491, "step": 1824 }, { "epoch": 0.19723332973089808, "grad_norm": 0.3530104160308838, "learning_rate": 8e-05, "loss": 1.6304, "step": 1825 }, { "epoch": 0.19734140278828488, "grad_norm": 0.34259703755378723, "learning_rate": 8e-05, "loss": 1.7162, "step": 1826 }, { "epoch": 0.19744947584567168, "grad_norm": 0.3622263967990875, "learning_rate": 8e-05, "loss": 1.8077, "step": 1827 }, { "epoch": 0.19755754890305846, "grad_norm": 0.39186692237854004, "learning_rate": 8e-05, "loss": 1.8964, "step": 1828 }, { "epoch": 0.19766562196044526, "grad_norm": 0.37032946944236755, "learning_rate": 8e-05, "loss": 1.7875, "step": 1829 }, { "epoch": 0.19777369501783204, "grad_norm": 0.3508176803588867, "learning_rate": 8e-05, "loss": 1.7005, "step": 1830 }, { "epoch": 0.19788176807521884, "grad_norm": 0.3399111330509186, "learning_rate": 8e-05, "loss": 1.6684, "step": 1831 }, { "epoch": 0.19798984113260565, "grad_norm": 0.37602466344833374, "learning_rate": 8e-05, "loss": 1.7446, "step": 1832 }, { "epoch": 0.19809791418999242, "grad_norm": 0.379396915435791, "learning_rate": 8e-05, "loss": 1.8423, "step": 1833 }, { "epoch": 0.19820598724737923, "grad_norm": 0.3804347813129425, "learning_rate": 8e-05, "loss": 1.5954, "step": 1834 }, { "epoch": 0.19831406030476603, "grad_norm": 0.3987898528575897, "learning_rate": 8e-05, "loss": 1.9973, "step": 1835 }, { "epoch": 0.1984221333621528, "grad_norm": 0.3639991879463196, "learning_rate": 8e-05, "loss": 1.7426, "step": 1836 }, { "epoch": 0.1985302064195396, "grad_norm": 0.3917749524116516, "learning_rate": 8e-05, "loss": 1.7536, "step": 1837 }, { "epoch": 0.19863827947692642, "grad_norm": 0.38116833567619324, "learning_rate": 8e-05, "loss": 1.8685, "step": 1838 }, { "epoch": 0.1987463525343132, "grad_norm": 0.34580197930336, "learning_rate": 8e-05, "loss": 1.7254, "step": 1839 }, { "epoch": 0.1988544255917, "grad_norm": 0.36838003993034363, "learning_rate": 8e-05, "loss": 1.955, "step": 1840 }, { "epoch": 0.19896249864908677, "grad_norm": 0.40240806341171265, "learning_rate": 8e-05, "loss": 1.8471, "step": 1841 }, { "epoch": 0.19907057170647358, "grad_norm": 0.3656959533691406, "learning_rate": 8e-05, "loss": 1.7889, "step": 1842 }, { "epoch": 0.19917864476386038, "grad_norm": 0.34446683526039124, "learning_rate": 8e-05, "loss": 1.6343, "step": 1843 }, { "epoch": 0.19928671782124716, "grad_norm": 0.3879225552082062, "learning_rate": 8e-05, "loss": 1.7978, "step": 1844 }, { "epoch": 0.19939479087863396, "grad_norm": 0.369751900434494, "learning_rate": 8e-05, "loss": 1.8596, "step": 1845 }, { "epoch": 0.19950286393602076, "grad_norm": 0.39010390639305115, "learning_rate": 8e-05, "loss": 1.8732, "step": 1846 }, { "epoch": 0.19961093699340754, "grad_norm": 0.3497869074344635, "learning_rate": 8e-05, "loss": 1.7282, "step": 1847 }, { "epoch": 0.19971901005079434, "grad_norm": 0.352588415145874, "learning_rate": 8e-05, "loss": 1.7256, "step": 1848 }, { "epoch": 0.19982708310818112, "grad_norm": 0.3593430817127228, "learning_rate": 8e-05, "loss": 1.7691, "step": 1849 }, { "epoch": 0.19993515616556792, "grad_norm": 0.3479427397251129, "learning_rate": 8e-05, "loss": 1.6938, "step": 1850 }, { "epoch": 0.20004322922295473, "grad_norm": 0.3976190686225891, "learning_rate": 8e-05, "loss": 1.8402, "step": 1851 }, { "epoch": 0.2001513022803415, "grad_norm": 0.36597055196762085, "learning_rate": 8e-05, "loss": 1.5968, "step": 1852 }, { "epoch": 0.2002593753377283, "grad_norm": 0.37614214420318604, "learning_rate": 8e-05, "loss": 1.8372, "step": 1853 }, { "epoch": 0.2003674483951151, "grad_norm": 0.36254647374153137, "learning_rate": 8e-05, "loss": 1.7885, "step": 1854 }, { "epoch": 0.2004755214525019, "grad_norm": 0.3588215112686157, "learning_rate": 8e-05, "loss": 1.7798, "step": 1855 }, { "epoch": 0.2005835945098887, "grad_norm": 0.33507752418518066, "learning_rate": 8e-05, "loss": 1.656, "step": 1856 }, { "epoch": 0.20069166756727547, "grad_norm": 0.3605280816555023, "learning_rate": 8e-05, "loss": 1.7644, "step": 1857 }, { "epoch": 0.20079974062466227, "grad_norm": 0.3604772984981537, "learning_rate": 8e-05, "loss": 1.7186, "step": 1858 }, { "epoch": 0.20090781368204907, "grad_norm": 0.38815072178840637, "learning_rate": 8e-05, "loss": 1.9593, "step": 1859 }, { "epoch": 0.20101588673943585, "grad_norm": 0.37124311923980713, "learning_rate": 8e-05, "loss": 1.7532, "step": 1860 }, { "epoch": 0.20112395979682265, "grad_norm": 0.34862321615219116, "learning_rate": 8e-05, "loss": 1.7519, "step": 1861 }, { "epoch": 0.20123203285420946, "grad_norm": 0.35736700892448425, "learning_rate": 8e-05, "loss": 1.7361, "step": 1862 }, { "epoch": 0.20134010591159623, "grad_norm": 0.3790404200553894, "learning_rate": 8e-05, "loss": 1.7438, "step": 1863 }, { "epoch": 0.20144817896898304, "grad_norm": 0.35792702436447144, "learning_rate": 8e-05, "loss": 1.8143, "step": 1864 }, { "epoch": 0.20155625202636981, "grad_norm": 0.3490006625652313, "learning_rate": 8e-05, "loss": 1.7612, "step": 1865 }, { "epoch": 0.20166432508375662, "grad_norm": 0.3531262278556824, "learning_rate": 8e-05, "loss": 1.6239, "step": 1866 }, { "epoch": 0.20177239814114342, "grad_norm": 0.38021284341812134, "learning_rate": 8e-05, "loss": 1.8541, "step": 1867 }, { "epoch": 0.2018804711985302, "grad_norm": 0.3699413537979126, "learning_rate": 8e-05, "loss": 1.92, "step": 1868 }, { "epoch": 0.201988544255917, "grad_norm": 0.3971005380153656, "learning_rate": 8e-05, "loss": 1.7421, "step": 1869 }, { "epoch": 0.2020966173133038, "grad_norm": 0.36012476682662964, "learning_rate": 8e-05, "loss": 1.7661, "step": 1870 }, { "epoch": 0.20220469037069058, "grad_norm": 0.36561453342437744, "learning_rate": 8e-05, "loss": 1.6984, "step": 1871 }, { "epoch": 0.20231276342807739, "grad_norm": 0.3798507750034332, "learning_rate": 8e-05, "loss": 1.8763, "step": 1872 }, { "epoch": 0.20242083648546416, "grad_norm": 0.3517743945121765, "learning_rate": 8e-05, "loss": 1.7786, "step": 1873 }, { "epoch": 0.20252890954285097, "grad_norm": 0.406732439994812, "learning_rate": 8e-05, "loss": 1.6092, "step": 1874 }, { "epoch": 0.20263698260023777, "grad_norm": 0.38151341676712036, "learning_rate": 8e-05, "loss": 1.8307, "step": 1875 }, { "epoch": 0.20274505565762455, "grad_norm": 0.3487846851348877, "learning_rate": 8e-05, "loss": 1.8087, "step": 1876 }, { "epoch": 0.20285312871501135, "grad_norm": 0.37762340903282166, "learning_rate": 8e-05, "loss": 1.7293, "step": 1877 }, { "epoch": 0.20296120177239815, "grad_norm": 0.386909544467926, "learning_rate": 8e-05, "loss": 1.7993, "step": 1878 }, { "epoch": 0.20306927482978493, "grad_norm": 0.3647140562534332, "learning_rate": 8e-05, "loss": 1.7843, "step": 1879 }, { "epoch": 0.20317734788717173, "grad_norm": 0.3647440969944, "learning_rate": 8e-05, "loss": 1.8405, "step": 1880 }, { "epoch": 0.2032854209445585, "grad_norm": 0.36391812562942505, "learning_rate": 8e-05, "loss": 1.7448, "step": 1881 }, { "epoch": 0.2033934940019453, "grad_norm": 0.384227991104126, "learning_rate": 8e-05, "loss": 1.8287, "step": 1882 }, { "epoch": 0.20350156705933212, "grad_norm": 0.36254239082336426, "learning_rate": 8e-05, "loss": 1.753, "step": 1883 }, { "epoch": 0.2036096401167189, "grad_norm": 0.398166686296463, "learning_rate": 8e-05, "loss": 1.9578, "step": 1884 }, { "epoch": 0.2037177131741057, "grad_norm": 0.3709028363227844, "learning_rate": 8e-05, "loss": 1.829, "step": 1885 }, { "epoch": 0.2038257862314925, "grad_norm": 0.35031962394714355, "learning_rate": 8e-05, "loss": 1.7362, "step": 1886 }, { "epoch": 0.20393385928887928, "grad_norm": 0.35780707001686096, "learning_rate": 8e-05, "loss": 1.8365, "step": 1887 }, { "epoch": 0.20404193234626608, "grad_norm": 0.3482782244682312, "learning_rate": 8e-05, "loss": 1.5845, "step": 1888 }, { "epoch": 0.20415000540365286, "grad_norm": 0.3618932366371155, "learning_rate": 8e-05, "loss": 1.7825, "step": 1889 }, { "epoch": 0.20425807846103966, "grad_norm": 0.35274481773376465, "learning_rate": 8e-05, "loss": 1.7753, "step": 1890 }, { "epoch": 0.20436615151842646, "grad_norm": 0.41224750876426697, "learning_rate": 8e-05, "loss": 1.784, "step": 1891 }, { "epoch": 0.20447422457581324, "grad_norm": 0.3973805606365204, "learning_rate": 8e-05, "loss": 1.872, "step": 1892 }, { "epoch": 0.20458229763320004, "grad_norm": 0.4106702208518982, "learning_rate": 8e-05, "loss": 1.9697, "step": 1893 }, { "epoch": 0.20469037069058685, "grad_norm": 0.38575512170791626, "learning_rate": 8e-05, "loss": 1.9009, "step": 1894 }, { "epoch": 0.20479844374797362, "grad_norm": 0.38011008501052856, "learning_rate": 8e-05, "loss": 1.7581, "step": 1895 }, { "epoch": 0.20490651680536043, "grad_norm": 0.3502074182033539, "learning_rate": 8e-05, "loss": 1.7413, "step": 1896 }, { "epoch": 0.2050145898627472, "grad_norm": 0.3682728409767151, "learning_rate": 8e-05, "loss": 1.8756, "step": 1897 }, { "epoch": 0.205122662920134, "grad_norm": 0.37342748045921326, "learning_rate": 8e-05, "loss": 1.6878, "step": 1898 }, { "epoch": 0.2052307359775208, "grad_norm": 0.365245521068573, "learning_rate": 8e-05, "loss": 1.6766, "step": 1899 }, { "epoch": 0.2053388090349076, "grad_norm": 0.3619349002838135, "learning_rate": 8e-05, "loss": 1.7115, "step": 1900 }, { "epoch": 0.2054468820922944, "grad_norm": 0.384347528219223, "learning_rate": 8e-05, "loss": 1.7359, "step": 1901 }, { "epoch": 0.2055549551496812, "grad_norm": 0.35721322894096375, "learning_rate": 8e-05, "loss": 1.7859, "step": 1902 }, { "epoch": 0.20566302820706797, "grad_norm": 0.3474719524383545, "learning_rate": 8e-05, "loss": 1.6841, "step": 1903 }, { "epoch": 0.20577110126445478, "grad_norm": 0.3767649829387665, "learning_rate": 8e-05, "loss": 1.8515, "step": 1904 }, { "epoch": 0.20587917432184155, "grad_norm": 0.34641364216804504, "learning_rate": 8e-05, "loss": 1.7236, "step": 1905 }, { "epoch": 0.20598724737922836, "grad_norm": 0.34312519431114197, "learning_rate": 8e-05, "loss": 1.6293, "step": 1906 }, { "epoch": 0.20609532043661516, "grad_norm": 0.3383251130580902, "learning_rate": 8e-05, "loss": 1.6503, "step": 1907 }, { "epoch": 0.20620339349400194, "grad_norm": 0.35973063111305237, "learning_rate": 8e-05, "loss": 1.7942, "step": 1908 }, { "epoch": 0.20631146655138874, "grad_norm": 0.3731791377067566, "learning_rate": 8e-05, "loss": 1.8547, "step": 1909 }, { "epoch": 0.20641953960877554, "grad_norm": 0.3464277684688568, "learning_rate": 8e-05, "loss": 1.7279, "step": 1910 }, { "epoch": 0.20652761266616232, "grad_norm": 0.3760031461715698, "learning_rate": 8e-05, "loss": 1.9154, "step": 1911 }, { "epoch": 0.20663568572354912, "grad_norm": 0.34432798624038696, "learning_rate": 8e-05, "loss": 1.4793, "step": 1912 }, { "epoch": 0.2067437587809359, "grad_norm": 0.3558083176612854, "learning_rate": 8e-05, "loss": 1.5359, "step": 1913 }, { "epoch": 0.2068518318383227, "grad_norm": 0.35777437686920166, "learning_rate": 8e-05, "loss": 1.747, "step": 1914 }, { "epoch": 0.2069599048957095, "grad_norm": 0.38074642419815063, "learning_rate": 8e-05, "loss": 1.8046, "step": 1915 }, { "epoch": 0.20706797795309628, "grad_norm": 0.34945687651634216, "learning_rate": 8e-05, "loss": 1.6817, "step": 1916 }, { "epoch": 0.2071760510104831, "grad_norm": 0.38633808493614197, "learning_rate": 8e-05, "loss": 1.7733, "step": 1917 }, { "epoch": 0.2072841240678699, "grad_norm": 0.40135735273361206, "learning_rate": 8e-05, "loss": 1.8261, "step": 1918 }, { "epoch": 0.20739219712525667, "grad_norm": 0.34329578280448914, "learning_rate": 8e-05, "loss": 1.7734, "step": 1919 }, { "epoch": 0.20750027018264347, "grad_norm": 0.38571685552597046, "learning_rate": 8e-05, "loss": 1.7238, "step": 1920 }, { "epoch": 0.20760834324003027, "grad_norm": 0.3572600483894348, "learning_rate": 8e-05, "loss": 1.6594, "step": 1921 }, { "epoch": 0.20771641629741705, "grad_norm": 0.35564470291137695, "learning_rate": 8e-05, "loss": 1.6699, "step": 1922 }, { "epoch": 0.20782448935480385, "grad_norm": 0.37275245785713196, "learning_rate": 8e-05, "loss": 1.8173, "step": 1923 }, { "epoch": 0.20793256241219063, "grad_norm": 0.39318376779556274, "learning_rate": 8e-05, "loss": 2.0055, "step": 1924 }, { "epoch": 0.20804063546957743, "grad_norm": 0.4063023328781128, "learning_rate": 8e-05, "loss": 1.8252, "step": 1925 }, { "epoch": 0.20814870852696424, "grad_norm": 0.3309997618198395, "learning_rate": 8e-05, "loss": 1.5401, "step": 1926 }, { "epoch": 0.20825678158435101, "grad_norm": 0.38798418641090393, "learning_rate": 8e-05, "loss": 1.771, "step": 1927 }, { "epoch": 0.20836485464173782, "grad_norm": 0.36899039149284363, "learning_rate": 8e-05, "loss": 1.7375, "step": 1928 }, { "epoch": 0.20847292769912462, "grad_norm": 0.36819252371788025, "learning_rate": 8e-05, "loss": 1.7692, "step": 1929 }, { "epoch": 0.2085810007565114, "grad_norm": 0.36590033769607544, "learning_rate": 8e-05, "loss": 1.823, "step": 1930 }, { "epoch": 0.2086890738138982, "grad_norm": 0.3395880162715912, "learning_rate": 8e-05, "loss": 1.7245, "step": 1931 }, { "epoch": 0.20879714687128498, "grad_norm": 0.3607260286808014, "learning_rate": 8e-05, "loss": 1.7543, "step": 1932 }, { "epoch": 0.20890521992867178, "grad_norm": 0.37475481629371643, "learning_rate": 8e-05, "loss": 1.7385, "step": 1933 }, { "epoch": 0.20901329298605859, "grad_norm": 0.42187097668647766, "learning_rate": 8e-05, "loss": 1.8316, "step": 1934 }, { "epoch": 0.20912136604344536, "grad_norm": 0.3628307580947876, "learning_rate": 8e-05, "loss": 1.9118, "step": 1935 }, { "epoch": 0.20922943910083217, "grad_norm": 0.3630234897136688, "learning_rate": 8e-05, "loss": 1.7933, "step": 1936 }, { "epoch": 0.20933751215821897, "grad_norm": 0.37966346740722656, "learning_rate": 8e-05, "loss": 1.7856, "step": 1937 }, { "epoch": 0.20944558521560575, "grad_norm": 0.4155793786048889, "learning_rate": 8e-05, "loss": 1.6914, "step": 1938 }, { "epoch": 0.20955365827299255, "grad_norm": 0.33756765723228455, "learning_rate": 8e-05, "loss": 1.5709, "step": 1939 }, { "epoch": 0.20966173133037933, "grad_norm": 0.373968243598938, "learning_rate": 8e-05, "loss": 1.7968, "step": 1940 }, { "epoch": 0.20976980438776613, "grad_norm": 0.3583020865917206, "learning_rate": 8e-05, "loss": 1.6967, "step": 1941 }, { "epoch": 0.20987787744515293, "grad_norm": 0.32911446690559387, "learning_rate": 8e-05, "loss": 1.5576, "step": 1942 }, { "epoch": 0.2099859505025397, "grad_norm": 0.3314995765686035, "learning_rate": 8e-05, "loss": 1.5973, "step": 1943 }, { "epoch": 0.2100940235599265, "grad_norm": 0.36070725321769714, "learning_rate": 8e-05, "loss": 1.6226, "step": 1944 }, { "epoch": 0.21020209661731332, "grad_norm": 0.4008345901966095, "learning_rate": 8e-05, "loss": 1.9592, "step": 1945 }, { "epoch": 0.2103101696747001, "grad_norm": 0.358479768037796, "learning_rate": 8e-05, "loss": 1.6871, "step": 1946 }, { "epoch": 0.2104182427320869, "grad_norm": 0.3629383444786072, "learning_rate": 8e-05, "loss": 1.6467, "step": 1947 }, { "epoch": 0.21052631578947367, "grad_norm": 0.4131973385810852, "learning_rate": 8e-05, "loss": 1.8011, "step": 1948 }, { "epoch": 0.21063438884686048, "grad_norm": 0.43551886081695557, "learning_rate": 8e-05, "loss": 1.9769, "step": 1949 }, { "epoch": 0.21074246190424728, "grad_norm": 0.3896951675415039, "learning_rate": 8e-05, "loss": 1.8156, "step": 1950 }, { "epoch": 0.21085053496163406, "grad_norm": 0.35681670904159546, "learning_rate": 8e-05, "loss": 1.7828, "step": 1951 }, { "epoch": 0.21095860801902086, "grad_norm": 0.40379849076271057, "learning_rate": 8e-05, "loss": 1.8967, "step": 1952 }, { "epoch": 0.21106668107640766, "grad_norm": 0.35377737879753113, "learning_rate": 8e-05, "loss": 1.5626, "step": 1953 }, { "epoch": 0.21117475413379444, "grad_norm": 0.3624890446662903, "learning_rate": 8e-05, "loss": 1.7137, "step": 1954 }, { "epoch": 0.21128282719118124, "grad_norm": 0.3532603979110718, "learning_rate": 8e-05, "loss": 1.4725, "step": 1955 }, { "epoch": 0.21139090024856802, "grad_norm": 0.3497554659843445, "learning_rate": 8e-05, "loss": 1.8049, "step": 1956 }, { "epoch": 0.21149897330595482, "grad_norm": 0.3954872190952301, "learning_rate": 8e-05, "loss": 1.8425, "step": 1957 }, { "epoch": 0.21160704636334163, "grad_norm": 0.3571591079235077, "learning_rate": 8e-05, "loss": 1.6638, "step": 1958 }, { "epoch": 0.2117151194207284, "grad_norm": 0.38056448101997375, "learning_rate": 8e-05, "loss": 1.777, "step": 1959 }, { "epoch": 0.2118231924781152, "grad_norm": 0.4163707196712494, "learning_rate": 8e-05, "loss": 1.7648, "step": 1960 }, { "epoch": 0.211931265535502, "grad_norm": 0.376526415348053, "learning_rate": 8e-05, "loss": 1.7368, "step": 1961 }, { "epoch": 0.2120393385928888, "grad_norm": 0.3950906991958618, "learning_rate": 8e-05, "loss": 1.795, "step": 1962 }, { "epoch": 0.2121474116502756, "grad_norm": 0.39837372303009033, "learning_rate": 8e-05, "loss": 1.8725, "step": 1963 }, { "epoch": 0.21225548470766237, "grad_norm": 0.41050609946250916, "learning_rate": 8e-05, "loss": 1.8765, "step": 1964 }, { "epoch": 0.21236355776504917, "grad_norm": 0.35144171118736267, "learning_rate": 8e-05, "loss": 1.7414, "step": 1965 }, { "epoch": 0.21247163082243598, "grad_norm": 0.3683284819126129, "learning_rate": 8e-05, "loss": 1.737, "step": 1966 }, { "epoch": 0.21257970387982275, "grad_norm": 0.38534876704216003, "learning_rate": 8e-05, "loss": 1.9354, "step": 1967 }, { "epoch": 0.21268777693720956, "grad_norm": 0.38828667998313904, "learning_rate": 8e-05, "loss": 1.8566, "step": 1968 }, { "epoch": 0.21279584999459636, "grad_norm": 0.4068783223628998, "learning_rate": 8e-05, "loss": 1.708, "step": 1969 }, { "epoch": 0.21290392305198313, "grad_norm": 0.3801378011703491, "learning_rate": 8e-05, "loss": 1.8297, "step": 1970 }, { "epoch": 0.21301199610936994, "grad_norm": 0.3865795135498047, "learning_rate": 8e-05, "loss": 1.7451, "step": 1971 }, { "epoch": 0.21312006916675671, "grad_norm": 0.357890248298645, "learning_rate": 8e-05, "loss": 1.7743, "step": 1972 }, { "epoch": 0.21322814222414352, "grad_norm": 0.3626871109008789, "learning_rate": 8e-05, "loss": 1.5904, "step": 1973 }, { "epoch": 0.21333621528153032, "grad_norm": 0.36060309410095215, "learning_rate": 8e-05, "loss": 1.7338, "step": 1974 }, { "epoch": 0.2134442883389171, "grad_norm": 0.3594733774662018, "learning_rate": 8e-05, "loss": 1.7579, "step": 1975 }, { "epoch": 0.2135523613963039, "grad_norm": 0.3680800795555115, "learning_rate": 8e-05, "loss": 1.6201, "step": 1976 }, { "epoch": 0.2136604344536907, "grad_norm": 0.3860113322734833, "learning_rate": 8e-05, "loss": 1.9166, "step": 1977 }, { "epoch": 0.21376850751107748, "grad_norm": 0.3709617853164673, "learning_rate": 8e-05, "loss": 1.8738, "step": 1978 }, { "epoch": 0.2138765805684643, "grad_norm": 0.37274497747421265, "learning_rate": 8e-05, "loss": 1.7343, "step": 1979 }, { "epoch": 0.21398465362585106, "grad_norm": 0.3914092481136322, "learning_rate": 8e-05, "loss": 1.8579, "step": 1980 }, { "epoch": 0.21409272668323787, "grad_norm": 0.40114396810531616, "learning_rate": 8e-05, "loss": 1.8042, "step": 1981 }, { "epoch": 0.21420079974062467, "grad_norm": 0.36645975708961487, "learning_rate": 8e-05, "loss": 1.7846, "step": 1982 }, { "epoch": 0.21430887279801145, "grad_norm": 0.36580151319503784, "learning_rate": 8e-05, "loss": 1.8357, "step": 1983 }, { "epoch": 0.21441694585539825, "grad_norm": 0.3511945605278015, "learning_rate": 8e-05, "loss": 1.6754, "step": 1984 }, { "epoch": 0.21452501891278505, "grad_norm": 0.3777111768722534, "learning_rate": 8e-05, "loss": 1.8998, "step": 1985 }, { "epoch": 0.21463309197017183, "grad_norm": 0.33652061223983765, "learning_rate": 8e-05, "loss": 1.7839, "step": 1986 }, { "epoch": 0.21474116502755863, "grad_norm": 0.3497997522354126, "learning_rate": 8e-05, "loss": 1.7363, "step": 1987 }, { "epoch": 0.2148492380849454, "grad_norm": 0.3635331392288208, "learning_rate": 8e-05, "loss": 1.7969, "step": 1988 }, { "epoch": 0.2149573111423322, "grad_norm": 0.3440324366092682, "learning_rate": 8e-05, "loss": 1.7119, "step": 1989 }, { "epoch": 0.21506538419971902, "grad_norm": 0.37369877099990845, "learning_rate": 8e-05, "loss": 1.7509, "step": 1990 }, { "epoch": 0.2151734572571058, "grad_norm": 0.35939815640449524, "learning_rate": 8e-05, "loss": 1.6149, "step": 1991 }, { "epoch": 0.2152815303144926, "grad_norm": 0.37687626481056213, "learning_rate": 8e-05, "loss": 1.747, "step": 1992 }, { "epoch": 0.2153896033718794, "grad_norm": 0.4100135862827301, "learning_rate": 8e-05, "loss": 1.8089, "step": 1993 }, { "epoch": 0.21549767642926618, "grad_norm": 0.3928000032901764, "learning_rate": 8e-05, "loss": 1.9531, "step": 1994 }, { "epoch": 0.21560574948665298, "grad_norm": 0.36357176303863525, "learning_rate": 8e-05, "loss": 1.8323, "step": 1995 }, { "epoch": 0.21571382254403976, "grad_norm": 0.40489187836647034, "learning_rate": 8e-05, "loss": 1.8864, "step": 1996 }, { "epoch": 0.21582189560142656, "grad_norm": 0.3513606786727905, "learning_rate": 8e-05, "loss": 1.7082, "step": 1997 }, { "epoch": 0.21592996865881336, "grad_norm": 0.3592759370803833, "learning_rate": 8e-05, "loss": 1.7024, "step": 1998 }, { "epoch": 0.21603804171620014, "grad_norm": 0.3832401931285858, "learning_rate": 8e-05, "loss": 1.8435, "step": 1999 }, { "epoch": 0.21614611477358694, "grad_norm": 0.35246819257736206, "learning_rate": 8e-05, "loss": 1.7322, "step": 2000 }, { "epoch": 0.21625418783097375, "grad_norm": 0.36942264437675476, "learning_rate": 8e-05, "loss": 1.8037, "step": 2001 }, { "epoch": 0.21636226088836052, "grad_norm": 0.3643299341201782, "learning_rate": 8e-05, "loss": 1.6774, "step": 2002 }, { "epoch": 0.21647033394574733, "grad_norm": 0.35385361313819885, "learning_rate": 8e-05, "loss": 1.6414, "step": 2003 }, { "epoch": 0.21657840700313413, "grad_norm": 0.36905983090400696, "learning_rate": 8e-05, "loss": 1.5822, "step": 2004 }, { "epoch": 0.2166864800605209, "grad_norm": 0.3492202162742615, "learning_rate": 8e-05, "loss": 1.6326, "step": 2005 }, { "epoch": 0.2167945531179077, "grad_norm": 0.4036939740180969, "learning_rate": 8e-05, "loss": 1.8589, "step": 2006 }, { "epoch": 0.2169026261752945, "grad_norm": 0.3529248833656311, "learning_rate": 8e-05, "loss": 1.7494, "step": 2007 }, { "epoch": 0.2170106992326813, "grad_norm": 0.3524319529533386, "learning_rate": 8e-05, "loss": 1.6436, "step": 2008 }, { "epoch": 0.2171187722900681, "grad_norm": 0.3989529311656952, "learning_rate": 8e-05, "loss": 1.7791, "step": 2009 }, { "epoch": 0.21722684534745487, "grad_norm": 0.3952706754207611, "learning_rate": 8e-05, "loss": 1.7964, "step": 2010 }, { "epoch": 0.21733491840484168, "grad_norm": 0.35971590876579285, "learning_rate": 8e-05, "loss": 1.6807, "step": 2011 }, { "epoch": 0.21744299146222848, "grad_norm": 0.4018203020095825, "learning_rate": 8e-05, "loss": 1.797, "step": 2012 }, { "epoch": 0.21755106451961526, "grad_norm": 0.38853293657302856, "learning_rate": 8e-05, "loss": 1.6688, "step": 2013 }, { "epoch": 0.21765913757700206, "grad_norm": 0.3590441644191742, "learning_rate": 8e-05, "loss": 1.4683, "step": 2014 }, { "epoch": 0.21776721063438884, "grad_norm": 0.37658265233039856, "learning_rate": 8e-05, "loss": 1.6729, "step": 2015 }, { "epoch": 0.21787528369177564, "grad_norm": 0.3557071387767792, "learning_rate": 8e-05, "loss": 1.6941, "step": 2016 }, { "epoch": 0.21798335674916244, "grad_norm": 0.37900692224502563, "learning_rate": 8e-05, "loss": 1.7138, "step": 2017 }, { "epoch": 0.21809142980654922, "grad_norm": 0.41085371375083923, "learning_rate": 8e-05, "loss": 1.8812, "step": 2018 }, { "epoch": 0.21819950286393602, "grad_norm": 0.3881458640098572, "learning_rate": 8e-05, "loss": 1.7512, "step": 2019 }, { "epoch": 0.21830757592132283, "grad_norm": 0.3675471246242523, "learning_rate": 8e-05, "loss": 1.769, "step": 2020 }, { "epoch": 0.2184156489787096, "grad_norm": 0.4067486524581909, "learning_rate": 8e-05, "loss": 1.7799, "step": 2021 }, { "epoch": 0.2185237220360964, "grad_norm": 0.3796285390853882, "learning_rate": 8e-05, "loss": 1.7437, "step": 2022 }, { "epoch": 0.21863179509348318, "grad_norm": 0.36941665410995483, "learning_rate": 8e-05, "loss": 1.6787, "step": 2023 }, { "epoch": 0.21873986815087, "grad_norm": 0.36870813369750977, "learning_rate": 8e-05, "loss": 1.7539, "step": 2024 }, { "epoch": 0.2188479412082568, "grad_norm": 0.3542405664920807, "learning_rate": 8e-05, "loss": 1.6646, "step": 2025 }, { "epoch": 0.21895601426564357, "grad_norm": 0.37803706526756287, "learning_rate": 8e-05, "loss": 1.8062, "step": 2026 }, { "epoch": 0.21906408732303037, "grad_norm": 0.4057185649871826, "learning_rate": 8e-05, "loss": 1.7243, "step": 2027 }, { "epoch": 0.21917216038041717, "grad_norm": 0.39966803789138794, "learning_rate": 8e-05, "loss": 1.801, "step": 2028 }, { "epoch": 0.21928023343780395, "grad_norm": 0.36467891931533813, "learning_rate": 8e-05, "loss": 1.8367, "step": 2029 }, { "epoch": 0.21938830649519075, "grad_norm": 0.4220462441444397, "learning_rate": 8e-05, "loss": 1.7095, "step": 2030 }, { "epoch": 0.21949637955257753, "grad_norm": 0.3874486982822418, "learning_rate": 8e-05, "loss": 1.8462, "step": 2031 }, { "epoch": 0.21960445260996433, "grad_norm": 0.35149145126342773, "learning_rate": 8e-05, "loss": 1.7776, "step": 2032 }, { "epoch": 0.21971252566735114, "grad_norm": 0.366610586643219, "learning_rate": 8e-05, "loss": 1.774, "step": 2033 }, { "epoch": 0.21982059872473791, "grad_norm": 0.37655380368232727, "learning_rate": 8e-05, "loss": 1.8611, "step": 2034 }, { "epoch": 0.21992867178212472, "grad_norm": 0.39081716537475586, "learning_rate": 8e-05, "loss": 1.8509, "step": 2035 }, { "epoch": 0.22003674483951152, "grad_norm": 0.36093536019325256, "learning_rate": 8e-05, "loss": 1.5533, "step": 2036 }, { "epoch": 0.2201448178968983, "grad_norm": 0.37873879075050354, "learning_rate": 8e-05, "loss": 1.8671, "step": 2037 }, { "epoch": 0.2202528909542851, "grad_norm": 0.35921892523765564, "learning_rate": 8e-05, "loss": 1.6938, "step": 2038 }, { "epoch": 0.22036096401167188, "grad_norm": 0.34384745359420776, "learning_rate": 8e-05, "loss": 1.5841, "step": 2039 }, { "epoch": 0.22046903706905868, "grad_norm": 0.383731871843338, "learning_rate": 8e-05, "loss": 1.7347, "step": 2040 }, { "epoch": 0.22057711012644549, "grad_norm": 0.3606813848018646, "learning_rate": 8e-05, "loss": 1.7684, "step": 2041 }, { "epoch": 0.22068518318383226, "grad_norm": 0.3706587851047516, "learning_rate": 8e-05, "loss": 1.7448, "step": 2042 }, { "epoch": 0.22079325624121907, "grad_norm": 0.39703646302223206, "learning_rate": 8e-05, "loss": 1.7732, "step": 2043 }, { "epoch": 0.22090132929860587, "grad_norm": 0.37300845980644226, "learning_rate": 8e-05, "loss": 1.8311, "step": 2044 }, { "epoch": 0.22100940235599265, "grad_norm": 0.3679048418998718, "learning_rate": 8e-05, "loss": 1.7147, "step": 2045 }, { "epoch": 0.22111747541337945, "grad_norm": 0.36237233877182007, "learning_rate": 8e-05, "loss": 1.8042, "step": 2046 }, { "epoch": 0.22122554847076623, "grad_norm": 0.38080283999443054, "learning_rate": 8e-05, "loss": 1.6903, "step": 2047 }, { "epoch": 0.22133362152815303, "grad_norm": 0.3606734573841095, "learning_rate": 8e-05, "loss": 1.6842, "step": 2048 }, { "epoch": 0.22144169458553983, "grad_norm": 0.36421850323677063, "learning_rate": 8e-05, "loss": 1.6836, "step": 2049 }, { "epoch": 0.2215497676429266, "grad_norm": 0.4062765836715698, "learning_rate": 8e-05, "loss": 1.8359, "step": 2050 }, { "epoch": 0.2216578407003134, "grad_norm": 0.4163103699684143, "learning_rate": 8e-05, "loss": 1.9322, "step": 2051 }, { "epoch": 0.22176591375770022, "grad_norm": 0.3605847656726837, "learning_rate": 8e-05, "loss": 1.6446, "step": 2052 }, { "epoch": 0.221873986815087, "grad_norm": 0.38698506355285645, "learning_rate": 8e-05, "loss": 1.8378, "step": 2053 }, { "epoch": 0.2219820598724738, "grad_norm": 0.3885916769504547, "learning_rate": 8e-05, "loss": 1.8099, "step": 2054 }, { "epoch": 0.22209013292986057, "grad_norm": 0.3732059895992279, "learning_rate": 8e-05, "loss": 1.6975, "step": 2055 }, { "epoch": 0.22219820598724738, "grad_norm": 0.4028705656528473, "learning_rate": 8e-05, "loss": 1.7729, "step": 2056 }, { "epoch": 0.22230627904463418, "grad_norm": 0.39209872484207153, "learning_rate": 8e-05, "loss": 1.9357, "step": 2057 }, { "epoch": 0.22241435210202096, "grad_norm": 0.3717482089996338, "learning_rate": 8e-05, "loss": 1.8335, "step": 2058 }, { "epoch": 0.22252242515940776, "grad_norm": 0.3872828483581543, "learning_rate": 8e-05, "loss": 1.852, "step": 2059 }, { "epoch": 0.22263049821679456, "grad_norm": 0.36990293860435486, "learning_rate": 8e-05, "loss": 1.6782, "step": 2060 }, { "epoch": 0.22273857127418134, "grad_norm": 0.3654635548591614, "learning_rate": 8e-05, "loss": 1.8468, "step": 2061 }, { "epoch": 0.22284664433156814, "grad_norm": 0.36631882190704346, "learning_rate": 8e-05, "loss": 1.9319, "step": 2062 }, { "epoch": 0.22295471738895492, "grad_norm": 0.37345147132873535, "learning_rate": 8e-05, "loss": 1.7308, "step": 2063 }, { "epoch": 0.22306279044634172, "grad_norm": 0.38322269916534424, "learning_rate": 8e-05, "loss": 1.8156, "step": 2064 }, { "epoch": 0.22317086350372853, "grad_norm": 0.38158226013183594, "learning_rate": 8e-05, "loss": 1.8166, "step": 2065 }, { "epoch": 0.2232789365611153, "grad_norm": 0.3726826608181, "learning_rate": 8e-05, "loss": 1.7048, "step": 2066 }, { "epoch": 0.2233870096185021, "grad_norm": 0.3899937570095062, "learning_rate": 8e-05, "loss": 1.8623, "step": 2067 }, { "epoch": 0.2234950826758889, "grad_norm": 0.3797368109226227, "learning_rate": 8e-05, "loss": 1.8981, "step": 2068 }, { "epoch": 0.2236031557332757, "grad_norm": 0.37010857462882996, "learning_rate": 8e-05, "loss": 1.7753, "step": 2069 }, { "epoch": 0.2237112287906625, "grad_norm": 0.36409837007522583, "learning_rate": 8e-05, "loss": 1.7197, "step": 2070 }, { "epoch": 0.22381930184804927, "grad_norm": 0.42100897431373596, "learning_rate": 8e-05, "loss": 1.8354, "step": 2071 }, { "epoch": 0.22392737490543607, "grad_norm": 0.3848303258419037, "learning_rate": 8e-05, "loss": 1.8157, "step": 2072 }, { "epoch": 0.22403544796282288, "grad_norm": 0.35955747961997986, "learning_rate": 8e-05, "loss": 1.701, "step": 2073 }, { "epoch": 0.22414352102020965, "grad_norm": 0.3381285071372986, "learning_rate": 8e-05, "loss": 1.6696, "step": 2074 }, { "epoch": 0.22425159407759646, "grad_norm": 0.4168195426464081, "learning_rate": 8e-05, "loss": 1.8523, "step": 2075 }, { "epoch": 0.22435966713498326, "grad_norm": 0.3455525040626526, "learning_rate": 8e-05, "loss": 1.5917, "step": 2076 }, { "epoch": 0.22446774019237004, "grad_norm": 0.3746519386768341, "learning_rate": 8e-05, "loss": 1.809, "step": 2077 }, { "epoch": 0.22457581324975684, "grad_norm": 0.3646126985549927, "learning_rate": 8e-05, "loss": 1.7883, "step": 2078 }, { "epoch": 0.22468388630714362, "grad_norm": 0.4021371006965637, "learning_rate": 8e-05, "loss": 1.8259, "step": 2079 }, { "epoch": 0.22479195936453042, "grad_norm": 0.3647554814815521, "learning_rate": 8e-05, "loss": 1.7899, "step": 2080 }, { "epoch": 0.22490003242191722, "grad_norm": 0.37920671701431274, "learning_rate": 8e-05, "loss": 1.7413, "step": 2081 }, { "epoch": 0.225008105479304, "grad_norm": 0.39572465419769287, "learning_rate": 8e-05, "loss": 1.7501, "step": 2082 }, { "epoch": 0.2251161785366908, "grad_norm": 0.4094925820827484, "learning_rate": 8e-05, "loss": 1.7248, "step": 2083 }, { "epoch": 0.2252242515940776, "grad_norm": 0.3848681151866913, "learning_rate": 8e-05, "loss": 1.6909, "step": 2084 }, { "epoch": 0.22533232465146438, "grad_norm": 0.38531455397605896, "learning_rate": 8e-05, "loss": 1.8568, "step": 2085 }, { "epoch": 0.2254403977088512, "grad_norm": 0.3507544994354248, "learning_rate": 8e-05, "loss": 1.8107, "step": 2086 }, { "epoch": 0.225548470766238, "grad_norm": 0.39098140597343445, "learning_rate": 8e-05, "loss": 1.8424, "step": 2087 }, { "epoch": 0.22565654382362477, "grad_norm": 0.36397454142570496, "learning_rate": 8e-05, "loss": 1.6309, "step": 2088 }, { "epoch": 0.22576461688101157, "grad_norm": 0.4234897196292877, "learning_rate": 8e-05, "loss": 2.0841, "step": 2089 }, { "epoch": 0.22587268993839835, "grad_norm": 0.3712233006954193, "learning_rate": 8e-05, "loss": 1.7417, "step": 2090 }, { "epoch": 0.22598076299578515, "grad_norm": 0.34451958537101746, "learning_rate": 8e-05, "loss": 1.7492, "step": 2091 }, { "epoch": 0.22608883605317195, "grad_norm": 0.38284558057785034, "learning_rate": 8e-05, "loss": 1.9053, "step": 2092 }, { "epoch": 0.22619690911055873, "grad_norm": 0.35918140411376953, "learning_rate": 8e-05, "loss": 1.7304, "step": 2093 }, { "epoch": 0.22630498216794553, "grad_norm": 0.4118683338165283, "learning_rate": 8e-05, "loss": 1.5915, "step": 2094 }, { "epoch": 0.22641305522533234, "grad_norm": 0.34408804774284363, "learning_rate": 8e-05, "loss": 1.6893, "step": 2095 }, { "epoch": 0.22652112828271911, "grad_norm": 0.37702298164367676, "learning_rate": 8e-05, "loss": 1.6662, "step": 2096 }, { "epoch": 0.22662920134010592, "grad_norm": 0.41421517729759216, "learning_rate": 8e-05, "loss": 2.0072, "step": 2097 }, { "epoch": 0.2267372743974927, "grad_norm": 0.3741128146648407, "learning_rate": 8e-05, "loss": 1.8918, "step": 2098 }, { "epoch": 0.2268453474548795, "grad_norm": 0.36649420857429504, "learning_rate": 8e-05, "loss": 1.7132, "step": 2099 }, { "epoch": 0.2269534205122663, "grad_norm": 0.34922367334365845, "learning_rate": 8e-05, "loss": 1.7296, "step": 2100 }, { "epoch": 0.22706149356965308, "grad_norm": 0.40428659319877625, "learning_rate": 8e-05, "loss": 1.6756, "step": 2101 }, { "epoch": 0.22716956662703988, "grad_norm": 0.33680659532546997, "learning_rate": 8e-05, "loss": 1.624, "step": 2102 }, { "epoch": 0.22727763968442669, "grad_norm": 0.3802899122238159, "learning_rate": 8e-05, "loss": 1.8264, "step": 2103 }, { "epoch": 0.22738571274181346, "grad_norm": 0.37044817209243774, "learning_rate": 8e-05, "loss": 1.7228, "step": 2104 }, { "epoch": 0.22749378579920027, "grad_norm": 0.3547819256782532, "learning_rate": 8e-05, "loss": 1.851, "step": 2105 }, { "epoch": 0.22760185885658704, "grad_norm": 0.36910563707351685, "learning_rate": 8e-05, "loss": 1.729, "step": 2106 }, { "epoch": 0.22770993191397385, "grad_norm": 0.4009215533733368, "learning_rate": 8e-05, "loss": 1.7693, "step": 2107 }, { "epoch": 0.22781800497136065, "grad_norm": 0.39027926325798035, "learning_rate": 8e-05, "loss": 1.859, "step": 2108 }, { "epoch": 0.22792607802874743, "grad_norm": 0.3891124427318573, "learning_rate": 8e-05, "loss": 1.8116, "step": 2109 }, { "epoch": 0.22803415108613423, "grad_norm": 0.3485018014907837, "learning_rate": 8e-05, "loss": 1.7495, "step": 2110 }, { "epoch": 0.22814222414352103, "grad_norm": 0.3762092590332031, "learning_rate": 8e-05, "loss": 1.7497, "step": 2111 }, { "epoch": 0.2282502972009078, "grad_norm": 0.3456081449985504, "learning_rate": 8e-05, "loss": 1.5313, "step": 2112 }, { "epoch": 0.2283583702582946, "grad_norm": 0.3693814277648926, "learning_rate": 8e-05, "loss": 1.6986, "step": 2113 }, { "epoch": 0.2284664433156814, "grad_norm": 0.34173548221588135, "learning_rate": 8e-05, "loss": 1.6831, "step": 2114 }, { "epoch": 0.2285745163730682, "grad_norm": 0.3535604476928711, "learning_rate": 8e-05, "loss": 1.7185, "step": 2115 }, { "epoch": 0.228682589430455, "grad_norm": 0.4323752224445343, "learning_rate": 8e-05, "loss": 1.8706, "step": 2116 }, { "epoch": 0.22879066248784177, "grad_norm": 0.3621761202812195, "learning_rate": 8e-05, "loss": 1.7453, "step": 2117 }, { "epoch": 0.22889873554522858, "grad_norm": 0.3651280999183655, "learning_rate": 8e-05, "loss": 1.8356, "step": 2118 }, { "epoch": 0.22900680860261538, "grad_norm": 0.36092856526374817, "learning_rate": 8e-05, "loss": 1.7377, "step": 2119 }, { "epoch": 0.22911488166000216, "grad_norm": 0.3604333996772766, "learning_rate": 8e-05, "loss": 1.7117, "step": 2120 }, { "epoch": 0.22922295471738896, "grad_norm": 0.37237662076950073, "learning_rate": 8e-05, "loss": 1.8143, "step": 2121 }, { "epoch": 0.22933102777477574, "grad_norm": 0.3683346211910248, "learning_rate": 8e-05, "loss": 1.8115, "step": 2122 }, { "epoch": 0.22943910083216254, "grad_norm": 0.37433743476867676, "learning_rate": 8e-05, "loss": 1.751, "step": 2123 }, { "epoch": 0.22954717388954934, "grad_norm": 0.36272990703582764, "learning_rate": 8e-05, "loss": 1.7943, "step": 2124 }, { "epoch": 0.22965524694693612, "grad_norm": 0.36922699213027954, "learning_rate": 8e-05, "loss": 1.7116, "step": 2125 }, { "epoch": 0.22976332000432292, "grad_norm": 0.3609287738800049, "learning_rate": 8e-05, "loss": 1.6693, "step": 2126 }, { "epoch": 0.22987139306170973, "grad_norm": 0.36926889419555664, "learning_rate": 8e-05, "loss": 1.6912, "step": 2127 }, { "epoch": 0.2299794661190965, "grad_norm": 0.379178911447525, "learning_rate": 8e-05, "loss": 1.6821, "step": 2128 }, { "epoch": 0.2300875391764833, "grad_norm": 0.3815479278564453, "learning_rate": 8e-05, "loss": 1.9305, "step": 2129 }, { "epoch": 0.23019561223387008, "grad_norm": 0.3472154140472412, "learning_rate": 8e-05, "loss": 1.7167, "step": 2130 }, { "epoch": 0.2303036852912569, "grad_norm": 0.400835782289505, "learning_rate": 8e-05, "loss": 1.9232, "step": 2131 }, { "epoch": 0.2304117583486437, "grad_norm": 0.35532715916633606, "learning_rate": 8e-05, "loss": 1.7423, "step": 2132 }, { "epoch": 0.23051983140603047, "grad_norm": 0.376740425825119, "learning_rate": 8e-05, "loss": 1.9009, "step": 2133 }, { "epoch": 0.23062790446341727, "grad_norm": 0.3551524877548218, "learning_rate": 8e-05, "loss": 1.7351, "step": 2134 }, { "epoch": 0.23073597752080408, "grad_norm": 0.3769424855709076, "learning_rate": 8e-05, "loss": 1.598, "step": 2135 }, { "epoch": 0.23084405057819085, "grad_norm": 0.3506765365600586, "learning_rate": 8e-05, "loss": 1.7097, "step": 2136 }, { "epoch": 0.23095212363557766, "grad_norm": 0.3650258183479309, "learning_rate": 8e-05, "loss": 1.7816, "step": 2137 }, { "epoch": 0.23106019669296443, "grad_norm": 0.4579632878303528, "learning_rate": 8e-05, "loss": 1.911, "step": 2138 }, { "epoch": 0.23116826975035124, "grad_norm": 0.3669975996017456, "learning_rate": 8e-05, "loss": 1.7625, "step": 2139 }, { "epoch": 0.23127634280773804, "grad_norm": 0.33697396516799927, "learning_rate": 8e-05, "loss": 1.643, "step": 2140 }, { "epoch": 0.23138441586512482, "grad_norm": 0.3470466434955597, "learning_rate": 8e-05, "loss": 1.7947, "step": 2141 }, { "epoch": 0.23149248892251162, "grad_norm": 0.43133723735809326, "learning_rate": 8e-05, "loss": 1.6014, "step": 2142 }, { "epoch": 0.23160056197989842, "grad_norm": 0.34392091631889343, "learning_rate": 8e-05, "loss": 1.5699, "step": 2143 }, { "epoch": 0.2317086350372852, "grad_norm": 0.39652612805366516, "learning_rate": 8e-05, "loss": 1.9443, "step": 2144 }, { "epoch": 0.231816708094672, "grad_norm": 0.5036067366600037, "learning_rate": 8e-05, "loss": 1.6863, "step": 2145 }, { "epoch": 0.23192478115205878, "grad_norm": 0.36674514412879944, "learning_rate": 8e-05, "loss": 1.612, "step": 2146 }, { "epoch": 0.23203285420944558, "grad_norm": 0.37338146567344666, "learning_rate": 8e-05, "loss": 1.8063, "step": 2147 }, { "epoch": 0.2321409272668324, "grad_norm": 0.36324188113212585, "learning_rate": 8e-05, "loss": 1.6686, "step": 2148 }, { "epoch": 0.23224900032421916, "grad_norm": 0.38042178750038147, "learning_rate": 8e-05, "loss": 1.8055, "step": 2149 }, { "epoch": 0.23235707338160597, "grad_norm": 0.37401559948921204, "learning_rate": 8e-05, "loss": 1.8218, "step": 2150 }, { "epoch": 0.23246514643899277, "grad_norm": 0.34630411863327026, "learning_rate": 8e-05, "loss": 1.6044, "step": 2151 }, { "epoch": 0.23257321949637955, "grad_norm": 0.39804714918136597, "learning_rate": 8e-05, "loss": 1.9013, "step": 2152 }, { "epoch": 0.23268129255376635, "grad_norm": 0.40547269582748413, "learning_rate": 8e-05, "loss": 1.6791, "step": 2153 }, { "epoch": 0.23278936561115313, "grad_norm": 0.3861345052719116, "learning_rate": 8e-05, "loss": 1.75, "step": 2154 }, { "epoch": 0.23289743866853993, "grad_norm": 0.46861591935157776, "learning_rate": 8e-05, "loss": 2.1237, "step": 2155 }, { "epoch": 0.23300551172592673, "grad_norm": 0.35631823539733887, "learning_rate": 8e-05, "loss": 1.8165, "step": 2156 }, { "epoch": 0.2331135847833135, "grad_norm": 0.3970722258090973, "learning_rate": 8e-05, "loss": 1.6968, "step": 2157 }, { "epoch": 0.2332216578407003, "grad_norm": 0.3695705533027649, "learning_rate": 8e-05, "loss": 1.6869, "step": 2158 }, { "epoch": 0.23332973089808712, "grad_norm": 0.37531137466430664, "learning_rate": 8e-05, "loss": 1.9258, "step": 2159 }, { "epoch": 0.2334378039554739, "grad_norm": 0.4071751832962036, "learning_rate": 8e-05, "loss": 1.7253, "step": 2160 }, { "epoch": 0.2335458770128607, "grad_norm": 0.3767264783382416, "learning_rate": 8e-05, "loss": 1.8196, "step": 2161 }, { "epoch": 0.23365395007024747, "grad_norm": 0.42254704236984253, "learning_rate": 8e-05, "loss": 1.8827, "step": 2162 }, { "epoch": 0.23376202312763428, "grad_norm": 0.377869576215744, "learning_rate": 8e-05, "loss": 1.8, "step": 2163 }, { "epoch": 0.23387009618502108, "grad_norm": 0.3648615777492523, "learning_rate": 8e-05, "loss": 1.7956, "step": 2164 }, { "epoch": 0.23397816924240786, "grad_norm": 0.38333049416542053, "learning_rate": 8e-05, "loss": 1.8643, "step": 2165 }, { "epoch": 0.23408624229979466, "grad_norm": 0.37670475244522095, "learning_rate": 8e-05, "loss": 1.6685, "step": 2166 }, { "epoch": 0.23419431535718147, "grad_norm": 0.36983415484428406, "learning_rate": 8e-05, "loss": 1.8155, "step": 2167 }, { "epoch": 0.23430238841456824, "grad_norm": 0.3667638599872589, "learning_rate": 8e-05, "loss": 1.728, "step": 2168 }, { "epoch": 0.23441046147195505, "grad_norm": 0.41168835759162903, "learning_rate": 8e-05, "loss": 1.7782, "step": 2169 }, { "epoch": 0.23451853452934185, "grad_norm": 0.35350337624549866, "learning_rate": 8e-05, "loss": 1.6438, "step": 2170 }, { "epoch": 0.23462660758672862, "grad_norm": 0.3681151270866394, "learning_rate": 8e-05, "loss": 1.7795, "step": 2171 }, { "epoch": 0.23473468064411543, "grad_norm": 0.34820547699928284, "learning_rate": 8e-05, "loss": 1.7479, "step": 2172 }, { "epoch": 0.2348427537015022, "grad_norm": 0.35955289006233215, "learning_rate": 8e-05, "loss": 1.7299, "step": 2173 }, { "epoch": 0.234950826758889, "grad_norm": 0.3621039092540741, "learning_rate": 8e-05, "loss": 1.6894, "step": 2174 }, { "epoch": 0.2350588998162758, "grad_norm": 0.3995386064052582, "learning_rate": 8e-05, "loss": 1.8909, "step": 2175 }, { "epoch": 0.2351669728736626, "grad_norm": 0.37898147106170654, "learning_rate": 8e-05, "loss": 1.7872, "step": 2176 }, { "epoch": 0.2352750459310494, "grad_norm": 0.3780369460582733, "learning_rate": 8e-05, "loss": 1.8547, "step": 2177 }, { "epoch": 0.2353831189884362, "grad_norm": 0.3730384111404419, "learning_rate": 8e-05, "loss": 1.786, "step": 2178 }, { "epoch": 0.23549119204582297, "grad_norm": 0.357271671295166, "learning_rate": 8e-05, "loss": 1.7511, "step": 2179 }, { "epoch": 0.23559926510320978, "grad_norm": 0.3849795162677765, "learning_rate": 8e-05, "loss": 1.7558, "step": 2180 }, { "epoch": 0.23570733816059655, "grad_norm": 0.35723817348480225, "learning_rate": 8e-05, "loss": 1.6932, "step": 2181 }, { "epoch": 0.23581541121798336, "grad_norm": 0.37327802181243896, "learning_rate": 8e-05, "loss": 1.9252, "step": 2182 }, { "epoch": 0.23592348427537016, "grad_norm": 0.3704034090042114, "learning_rate": 8e-05, "loss": 1.7103, "step": 2183 }, { "epoch": 0.23603155733275694, "grad_norm": 0.375997930765152, "learning_rate": 8e-05, "loss": 1.844, "step": 2184 }, { "epoch": 0.23613963039014374, "grad_norm": 0.35905808210372925, "learning_rate": 8e-05, "loss": 1.6496, "step": 2185 }, { "epoch": 0.23624770344753054, "grad_norm": 0.3800273835659027, "learning_rate": 8e-05, "loss": 1.8063, "step": 2186 }, { "epoch": 0.23635577650491732, "grad_norm": 0.3805461823940277, "learning_rate": 8e-05, "loss": 1.8344, "step": 2187 }, { "epoch": 0.23646384956230412, "grad_norm": 0.36678841710090637, "learning_rate": 8e-05, "loss": 1.847, "step": 2188 }, { "epoch": 0.2365719226196909, "grad_norm": 0.37296822667121887, "learning_rate": 8e-05, "loss": 1.7978, "step": 2189 }, { "epoch": 0.2366799956770777, "grad_norm": 0.3672078847885132, "learning_rate": 8e-05, "loss": 1.8035, "step": 2190 }, { "epoch": 0.2367880687344645, "grad_norm": 0.35733985900878906, "learning_rate": 8e-05, "loss": 1.809, "step": 2191 }, { "epoch": 0.23689614179185128, "grad_norm": 0.3379526138305664, "learning_rate": 8e-05, "loss": 1.6965, "step": 2192 }, { "epoch": 0.2370042148492381, "grad_norm": 0.3835746943950653, "learning_rate": 8e-05, "loss": 1.8199, "step": 2193 }, { "epoch": 0.2371122879066249, "grad_norm": 0.49968793988227844, "learning_rate": 8e-05, "loss": 1.8153, "step": 2194 }, { "epoch": 0.23722036096401167, "grad_norm": 0.3420034945011139, "learning_rate": 8e-05, "loss": 1.6382, "step": 2195 }, { "epoch": 0.23732843402139847, "grad_norm": 0.3732185959815979, "learning_rate": 8e-05, "loss": 1.6961, "step": 2196 }, { "epoch": 0.23743650707878525, "grad_norm": 0.3790595531463623, "learning_rate": 8e-05, "loss": 1.6562, "step": 2197 }, { "epoch": 0.23754458013617205, "grad_norm": 0.3528604507446289, "learning_rate": 8e-05, "loss": 1.7056, "step": 2198 }, { "epoch": 0.23765265319355885, "grad_norm": 0.3715028762817383, "learning_rate": 8e-05, "loss": 1.627, "step": 2199 }, { "epoch": 0.23776072625094563, "grad_norm": 0.41230669617652893, "learning_rate": 8e-05, "loss": 1.901, "step": 2200 }, { "epoch": 0.23786879930833243, "grad_norm": 0.3810524046421051, "learning_rate": 8e-05, "loss": 1.8767, "step": 2201 }, { "epoch": 0.23797687236571924, "grad_norm": 0.37211164832115173, "learning_rate": 8e-05, "loss": 1.7414, "step": 2202 }, { "epoch": 0.23808494542310601, "grad_norm": 0.39589688181877136, "learning_rate": 8e-05, "loss": 1.7844, "step": 2203 }, { "epoch": 0.23819301848049282, "grad_norm": 0.3493586480617523, "learning_rate": 8e-05, "loss": 1.6978, "step": 2204 }, { "epoch": 0.2383010915378796, "grad_norm": 0.35826364159584045, "learning_rate": 8e-05, "loss": 1.6375, "step": 2205 }, { "epoch": 0.2384091645952664, "grad_norm": 0.3929729759693146, "learning_rate": 8e-05, "loss": 1.8507, "step": 2206 }, { "epoch": 0.2385172376526532, "grad_norm": 0.38978311419487, "learning_rate": 8e-05, "loss": 1.8152, "step": 2207 }, { "epoch": 0.23862531071003998, "grad_norm": 0.3733316957950592, "learning_rate": 8e-05, "loss": 1.7432, "step": 2208 }, { "epoch": 0.23873338376742678, "grad_norm": 0.3959824740886688, "learning_rate": 8e-05, "loss": 1.7615, "step": 2209 }, { "epoch": 0.23884145682481359, "grad_norm": 0.3888351023197174, "learning_rate": 8e-05, "loss": 1.8095, "step": 2210 }, { "epoch": 0.23894952988220036, "grad_norm": 0.38295966386795044, "learning_rate": 8e-05, "loss": 1.8071, "step": 2211 }, { "epoch": 0.23905760293958717, "grad_norm": 0.3820528984069824, "learning_rate": 8e-05, "loss": 1.6692, "step": 2212 }, { "epoch": 0.23916567599697394, "grad_norm": 0.3936896026134491, "learning_rate": 8e-05, "loss": 1.7392, "step": 2213 }, { "epoch": 0.23927374905436075, "grad_norm": 0.3367980420589447, "learning_rate": 8e-05, "loss": 1.6708, "step": 2214 }, { "epoch": 0.23938182211174755, "grad_norm": 0.38799363374710083, "learning_rate": 8e-05, "loss": 1.6317, "step": 2215 }, { "epoch": 0.23948989516913433, "grad_norm": 0.3720891773700714, "learning_rate": 8e-05, "loss": 1.7849, "step": 2216 }, { "epoch": 0.23959796822652113, "grad_norm": 0.34050485491752625, "learning_rate": 8e-05, "loss": 1.6556, "step": 2217 }, { "epoch": 0.23970604128390793, "grad_norm": 0.44663381576538086, "learning_rate": 8e-05, "loss": 1.9447, "step": 2218 }, { "epoch": 0.2398141143412947, "grad_norm": 0.3856361210346222, "learning_rate": 8e-05, "loss": 1.8881, "step": 2219 }, { "epoch": 0.2399221873986815, "grad_norm": 0.37233471870422363, "learning_rate": 8e-05, "loss": 1.7531, "step": 2220 }, { "epoch": 0.2400302604560683, "grad_norm": 0.36970770359039307, "learning_rate": 8e-05, "loss": 1.7351, "step": 2221 }, { "epoch": 0.2401383335134551, "grad_norm": 0.3888806104660034, "learning_rate": 8e-05, "loss": 1.6986, "step": 2222 }, { "epoch": 0.2402464065708419, "grad_norm": 0.3729874789714813, "learning_rate": 8e-05, "loss": 1.7228, "step": 2223 }, { "epoch": 0.24035447962822867, "grad_norm": 0.38584765791893005, "learning_rate": 8e-05, "loss": 1.7654, "step": 2224 }, { "epoch": 0.24046255268561548, "grad_norm": 0.39203691482543945, "learning_rate": 8e-05, "loss": 1.8621, "step": 2225 }, { "epoch": 0.24057062574300228, "grad_norm": 0.40381258726119995, "learning_rate": 8e-05, "loss": 1.9093, "step": 2226 }, { "epoch": 0.24067869880038906, "grad_norm": 0.3886108100414276, "learning_rate": 8e-05, "loss": 1.6487, "step": 2227 }, { "epoch": 0.24078677185777586, "grad_norm": 0.37426671385765076, "learning_rate": 8e-05, "loss": 1.7012, "step": 2228 }, { "epoch": 0.24089484491516264, "grad_norm": 0.35715681314468384, "learning_rate": 8e-05, "loss": 1.8178, "step": 2229 }, { "epoch": 0.24100291797254944, "grad_norm": 0.3785596191883087, "learning_rate": 8e-05, "loss": 1.7912, "step": 2230 }, { "epoch": 0.24111099102993624, "grad_norm": 0.4022001028060913, "learning_rate": 8e-05, "loss": 1.884, "step": 2231 }, { "epoch": 0.24121906408732302, "grad_norm": 0.38375037908554077, "learning_rate": 8e-05, "loss": 1.7541, "step": 2232 }, { "epoch": 0.24132713714470982, "grad_norm": 0.40449509024620056, "learning_rate": 8e-05, "loss": 1.8424, "step": 2233 }, { "epoch": 0.24143521020209663, "grad_norm": 0.3777287006378174, "learning_rate": 8e-05, "loss": 1.8728, "step": 2234 }, { "epoch": 0.2415432832594834, "grad_norm": 0.36847957968711853, "learning_rate": 8e-05, "loss": 1.8695, "step": 2235 }, { "epoch": 0.2416513563168702, "grad_norm": 0.39545324444770813, "learning_rate": 8e-05, "loss": 1.8409, "step": 2236 }, { "epoch": 0.24175942937425698, "grad_norm": 0.3706943988800049, "learning_rate": 8e-05, "loss": 1.7777, "step": 2237 }, { "epoch": 0.2418675024316438, "grad_norm": 0.3927188515663147, "learning_rate": 8e-05, "loss": 1.8033, "step": 2238 }, { "epoch": 0.2419755754890306, "grad_norm": 0.37358346581459045, "learning_rate": 8e-05, "loss": 1.8419, "step": 2239 }, { "epoch": 0.24208364854641737, "grad_norm": 0.3923817276954651, "learning_rate": 8e-05, "loss": 1.8291, "step": 2240 }, { "epoch": 0.24219172160380417, "grad_norm": 0.3883258104324341, "learning_rate": 8e-05, "loss": 1.7932, "step": 2241 }, { "epoch": 0.24229979466119098, "grad_norm": 0.3817564845085144, "learning_rate": 8e-05, "loss": 1.6676, "step": 2242 }, { "epoch": 0.24240786771857775, "grad_norm": 0.4428686201572418, "learning_rate": 8e-05, "loss": 1.9983, "step": 2243 }, { "epoch": 0.24251594077596456, "grad_norm": 0.3858909606933594, "learning_rate": 8e-05, "loss": 1.776, "step": 2244 }, { "epoch": 0.24262401383335133, "grad_norm": 0.3450782895088196, "learning_rate": 8e-05, "loss": 1.483, "step": 2245 }, { "epoch": 0.24273208689073814, "grad_norm": 0.40456026792526245, "learning_rate": 8e-05, "loss": 1.7666, "step": 2246 }, { "epoch": 0.24284015994812494, "grad_norm": 0.3651934862136841, "learning_rate": 8e-05, "loss": 1.4756, "step": 2247 }, { "epoch": 0.24294823300551172, "grad_norm": 0.3923119306564331, "learning_rate": 8e-05, "loss": 1.7053, "step": 2248 }, { "epoch": 0.24305630606289852, "grad_norm": 0.38456419110298157, "learning_rate": 8e-05, "loss": 1.9331, "step": 2249 }, { "epoch": 0.24316437912028532, "grad_norm": 0.38729560375213623, "learning_rate": 8e-05, "loss": 1.7855, "step": 2250 }, { "epoch": 0.2432724521776721, "grad_norm": 0.38927167654037476, "learning_rate": 8e-05, "loss": 1.464, "step": 2251 }, { "epoch": 0.2433805252350589, "grad_norm": 0.3939606547355652, "learning_rate": 8e-05, "loss": 1.8651, "step": 2252 }, { "epoch": 0.2434885982924457, "grad_norm": 0.3740518093109131, "learning_rate": 8e-05, "loss": 1.701, "step": 2253 }, { "epoch": 0.24359667134983248, "grad_norm": 0.3723231852054596, "learning_rate": 8e-05, "loss": 1.8437, "step": 2254 }, { "epoch": 0.2437047444072193, "grad_norm": 0.36414843797683716, "learning_rate": 8e-05, "loss": 1.8114, "step": 2255 }, { "epoch": 0.24381281746460606, "grad_norm": 0.3751830756664276, "learning_rate": 8e-05, "loss": 1.7811, "step": 2256 }, { "epoch": 0.24392089052199287, "grad_norm": 0.3694900870323181, "learning_rate": 8e-05, "loss": 1.6954, "step": 2257 }, { "epoch": 0.24402896357937967, "grad_norm": 0.3713975250720978, "learning_rate": 8e-05, "loss": 1.5858, "step": 2258 }, { "epoch": 0.24413703663676645, "grad_norm": 0.3691284954547882, "learning_rate": 8e-05, "loss": 1.7574, "step": 2259 }, { "epoch": 0.24424510969415325, "grad_norm": 0.438029944896698, "learning_rate": 8e-05, "loss": 1.8166, "step": 2260 }, { "epoch": 0.24435318275154005, "grad_norm": 0.3459267318248749, "learning_rate": 8e-05, "loss": 1.5822, "step": 2261 }, { "epoch": 0.24446125580892683, "grad_norm": 0.3764677345752716, "learning_rate": 8e-05, "loss": 1.6692, "step": 2262 }, { "epoch": 0.24456932886631363, "grad_norm": 0.38184359669685364, "learning_rate": 8e-05, "loss": 1.8978, "step": 2263 }, { "epoch": 0.2446774019237004, "grad_norm": 0.3612769544124603, "learning_rate": 8e-05, "loss": 1.8164, "step": 2264 }, { "epoch": 0.24478547498108721, "grad_norm": 0.37051650881767273, "learning_rate": 8e-05, "loss": 1.6602, "step": 2265 }, { "epoch": 0.24489354803847402, "grad_norm": 0.3863441050052643, "learning_rate": 8e-05, "loss": 1.8059, "step": 2266 }, { "epoch": 0.2450016210958608, "grad_norm": 0.3679267168045044, "learning_rate": 8e-05, "loss": 1.6059, "step": 2267 }, { "epoch": 0.2451096941532476, "grad_norm": 0.36337706446647644, "learning_rate": 8e-05, "loss": 1.5827, "step": 2268 }, { "epoch": 0.2452177672106344, "grad_norm": 0.4200214445590973, "learning_rate": 8e-05, "loss": 1.771, "step": 2269 }, { "epoch": 0.24532584026802118, "grad_norm": 0.38746392726898193, "learning_rate": 8e-05, "loss": 1.8312, "step": 2270 }, { "epoch": 0.24543391332540798, "grad_norm": 0.36637184023857117, "learning_rate": 8e-05, "loss": 1.7339, "step": 2271 }, { "epoch": 0.24554198638279476, "grad_norm": 0.39609017968177795, "learning_rate": 8e-05, "loss": 1.8369, "step": 2272 }, { "epoch": 0.24565005944018156, "grad_norm": 0.3779478967189789, "learning_rate": 8e-05, "loss": 1.8543, "step": 2273 }, { "epoch": 0.24575813249756837, "grad_norm": 0.4619922637939453, "learning_rate": 8e-05, "loss": 1.9894, "step": 2274 }, { "epoch": 0.24586620555495514, "grad_norm": 0.3626987040042877, "learning_rate": 8e-05, "loss": 1.7789, "step": 2275 }, { "epoch": 0.24597427861234195, "grad_norm": 0.3633941411972046, "learning_rate": 8e-05, "loss": 1.7663, "step": 2276 }, { "epoch": 0.24608235166972875, "grad_norm": 0.33208996057510376, "learning_rate": 8e-05, "loss": 1.6041, "step": 2277 }, { "epoch": 0.24619042472711553, "grad_norm": 0.3755781054496765, "learning_rate": 8e-05, "loss": 1.7178, "step": 2278 }, { "epoch": 0.24629849778450233, "grad_norm": 0.404191255569458, "learning_rate": 8e-05, "loss": 1.574, "step": 2279 }, { "epoch": 0.2464065708418891, "grad_norm": 0.3851582407951355, "learning_rate": 8e-05, "loss": 1.6624, "step": 2280 }, { "epoch": 0.2465146438992759, "grad_norm": 0.381092369556427, "learning_rate": 8e-05, "loss": 1.5647, "step": 2281 }, { "epoch": 0.2466227169566627, "grad_norm": 0.36779624223709106, "learning_rate": 8e-05, "loss": 1.7416, "step": 2282 }, { "epoch": 0.2467307900140495, "grad_norm": 0.369118869304657, "learning_rate": 8e-05, "loss": 1.8087, "step": 2283 }, { "epoch": 0.2468388630714363, "grad_norm": 0.36656805872917175, "learning_rate": 8e-05, "loss": 1.7392, "step": 2284 }, { "epoch": 0.2469469361288231, "grad_norm": 0.3935786187648773, "learning_rate": 8e-05, "loss": 1.9369, "step": 2285 }, { "epoch": 0.24705500918620987, "grad_norm": 0.37046152353286743, "learning_rate": 8e-05, "loss": 1.8025, "step": 2286 }, { "epoch": 0.24716308224359668, "grad_norm": 0.3597521185874939, "learning_rate": 8e-05, "loss": 1.653, "step": 2287 }, { "epoch": 0.24727115530098345, "grad_norm": 0.3555164635181427, "learning_rate": 8e-05, "loss": 1.7164, "step": 2288 }, { "epoch": 0.24737922835837026, "grad_norm": 0.368020623922348, "learning_rate": 8e-05, "loss": 1.8372, "step": 2289 }, { "epoch": 0.24748730141575706, "grad_norm": 0.3481471836566925, "learning_rate": 8e-05, "loss": 1.4505, "step": 2290 }, { "epoch": 0.24759537447314384, "grad_norm": 0.38928261399269104, "learning_rate": 8e-05, "loss": 1.8588, "step": 2291 }, { "epoch": 0.24770344753053064, "grad_norm": 0.3435552716255188, "learning_rate": 8e-05, "loss": 1.7479, "step": 2292 }, { "epoch": 0.24781152058791744, "grad_norm": 0.36094003915786743, "learning_rate": 8e-05, "loss": 1.8201, "step": 2293 }, { "epoch": 0.24791959364530422, "grad_norm": 0.36277252435684204, "learning_rate": 8e-05, "loss": 1.7336, "step": 2294 }, { "epoch": 0.24802766670269102, "grad_norm": 0.4144380986690521, "learning_rate": 8e-05, "loss": 1.7447, "step": 2295 }, { "epoch": 0.2481357397600778, "grad_norm": 0.35916078090667725, "learning_rate": 8e-05, "loss": 1.7085, "step": 2296 }, { "epoch": 0.2482438128174646, "grad_norm": 0.43701934814453125, "learning_rate": 8e-05, "loss": 1.6971, "step": 2297 }, { "epoch": 0.2483518858748514, "grad_norm": 0.3719756603240967, "learning_rate": 8e-05, "loss": 1.6979, "step": 2298 }, { "epoch": 0.24845995893223818, "grad_norm": 0.36405685544013977, "learning_rate": 8e-05, "loss": 1.7976, "step": 2299 }, { "epoch": 0.248568031989625, "grad_norm": 0.3792782425880432, "learning_rate": 8e-05, "loss": 1.826, "step": 2300 }, { "epoch": 0.2486761050470118, "grad_norm": 0.36561650037765503, "learning_rate": 8e-05, "loss": 1.7456, "step": 2301 }, { "epoch": 0.24878417810439857, "grad_norm": 0.3584391474723816, "learning_rate": 8e-05, "loss": 1.7197, "step": 2302 }, { "epoch": 0.24889225116178537, "grad_norm": 0.36541834473609924, "learning_rate": 8e-05, "loss": 1.7175, "step": 2303 }, { "epoch": 0.24900032421917215, "grad_norm": 0.3484187126159668, "learning_rate": 8e-05, "loss": 1.6182, "step": 2304 }, { "epoch": 0.24910839727655895, "grad_norm": 0.3748144805431366, "learning_rate": 8e-05, "loss": 1.8136, "step": 2305 }, { "epoch": 0.24921647033394576, "grad_norm": 0.3921423554420471, "learning_rate": 8e-05, "loss": 1.8872, "step": 2306 }, { "epoch": 0.24932454339133253, "grad_norm": 0.3733462989330292, "learning_rate": 8e-05, "loss": 1.6957, "step": 2307 }, { "epoch": 0.24943261644871934, "grad_norm": 0.367025226354599, "learning_rate": 8e-05, "loss": 1.7848, "step": 2308 }, { "epoch": 0.24954068950610614, "grad_norm": 0.44055962562561035, "learning_rate": 8e-05, "loss": 1.875, "step": 2309 }, { "epoch": 0.24964876256349292, "grad_norm": 0.37285923957824707, "learning_rate": 8e-05, "loss": 1.8193, "step": 2310 }, { "epoch": 0.24975683562087972, "grad_norm": 0.37885582447052, "learning_rate": 8e-05, "loss": 1.6128, "step": 2311 }, { "epoch": 0.2498649086782665, "grad_norm": 0.34981241822242737, "learning_rate": 8e-05, "loss": 1.5928, "step": 2312 }, { "epoch": 0.2499729817356533, "grad_norm": 0.3603290319442749, "learning_rate": 8e-05, "loss": 1.5015, "step": 2313 }, { "epoch": 0.2500810547930401, "grad_norm": 0.39479121565818787, "learning_rate": 8e-05, "loss": 1.8576, "step": 2314 }, { "epoch": 0.2501891278504269, "grad_norm": 0.37982019782066345, "learning_rate": 8e-05, "loss": 1.8667, "step": 2315 }, { "epoch": 0.25029720090781366, "grad_norm": 0.37069183588027954, "learning_rate": 8e-05, "loss": 1.7018, "step": 2316 }, { "epoch": 0.25040527396520046, "grad_norm": 0.3424277901649475, "learning_rate": 8e-05, "loss": 1.6891, "step": 2317 }, { "epoch": 0.25051334702258726, "grad_norm": 0.3897865116596222, "learning_rate": 8e-05, "loss": 1.9812, "step": 2318 }, { "epoch": 0.25062142007997407, "grad_norm": 0.359379380941391, "learning_rate": 8e-05, "loss": 1.6897, "step": 2319 }, { "epoch": 0.25072949313736087, "grad_norm": 0.3432386815547943, "learning_rate": 8e-05, "loss": 1.6447, "step": 2320 }, { "epoch": 0.2508375661947477, "grad_norm": 0.40121743083000183, "learning_rate": 8e-05, "loss": 1.8559, "step": 2321 }, { "epoch": 0.2509456392521344, "grad_norm": 0.37241867184638977, "learning_rate": 8e-05, "loss": 1.7152, "step": 2322 }, { "epoch": 0.2510537123095212, "grad_norm": 0.377501517534256, "learning_rate": 8e-05, "loss": 1.8125, "step": 2323 }, { "epoch": 0.25116178536690803, "grad_norm": 0.4114398658275604, "learning_rate": 8e-05, "loss": 1.8548, "step": 2324 }, { "epoch": 0.25126985842429483, "grad_norm": 0.3749444782733917, "learning_rate": 8e-05, "loss": 1.7083, "step": 2325 }, { "epoch": 0.25137793148168164, "grad_norm": 0.36130374670028687, "learning_rate": 8e-05, "loss": 1.7288, "step": 2326 }, { "epoch": 0.2514860045390684, "grad_norm": 0.3502368628978729, "learning_rate": 8e-05, "loss": 1.6863, "step": 2327 }, { "epoch": 0.2515940775964552, "grad_norm": 0.3810741901397705, "learning_rate": 8e-05, "loss": 1.6337, "step": 2328 }, { "epoch": 0.251702150653842, "grad_norm": 0.38377493619918823, "learning_rate": 8e-05, "loss": 1.7772, "step": 2329 }, { "epoch": 0.2518102237112288, "grad_norm": 0.411726176738739, "learning_rate": 8e-05, "loss": 1.9209, "step": 2330 }, { "epoch": 0.2519182967686156, "grad_norm": 0.3568357229232788, "learning_rate": 8e-05, "loss": 1.6508, "step": 2331 }, { "epoch": 0.25202636982600235, "grad_norm": 0.3556637763977051, "learning_rate": 8e-05, "loss": 1.6868, "step": 2332 }, { "epoch": 0.25213444288338915, "grad_norm": 0.3710155189037323, "learning_rate": 8e-05, "loss": 1.6579, "step": 2333 }, { "epoch": 0.25224251594077596, "grad_norm": 0.42000019550323486, "learning_rate": 8e-05, "loss": 2.0026, "step": 2334 }, { "epoch": 0.25235058899816276, "grad_norm": 0.4007807970046997, "learning_rate": 8e-05, "loss": 1.9122, "step": 2335 }, { "epoch": 0.25245866205554957, "grad_norm": 0.43295976519584656, "learning_rate": 8e-05, "loss": 1.6588, "step": 2336 }, { "epoch": 0.25256673511293637, "grad_norm": 0.39074695110321045, "learning_rate": 8e-05, "loss": 1.8649, "step": 2337 }, { "epoch": 0.2526748081703231, "grad_norm": 0.37223386764526367, "learning_rate": 8e-05, "loss": 1.5521, "step": 2338 }, { "epoch": 0.2527828812277099, "grad_norm": 0.36089980602264404, "learning_rate": 8e-05, "loss": 1.7315, "step": 2339 }, { "epoch": 0.2528909542850967, "grad_norm": 0.36689791083335876, "learning_rate": 8e-05, "loss": 1.7525, "step": 2340 }, { "epoch": 0.25299902734248353, "grad_norm": 0.3840523958206177, "learning_rate": 8e-05, "loss": 1.7713, "step": 2341 }, { "epoch": 0.25310710039987033, "grad_norm": 0.3728722035884857, "learning_rate": 8e-05, "loss": 1.7755, "step": 2342 }, { "epoch": 0.2532151734572571, "grad_norm": 0.396951287984848, "learning_rate": 8e-05, "loss": 1.8, "step": 2343 }, { "epoch": 0.2533232465146439, "grad_norm": 0.35852232575416565, "learning_rate": 8e-05, "loss": 1.7579, "step": 2344 }, { "epoch": 0.2534313195720307, "grad_norm": 0.35360392928123474, "learning_rate": 8e-05, "loss": 1.7129, "step": 2345 }, { "epoch": 0.2535393926294175, "grad_norm": 0.35659259557724, "learning_rate": 8e-05, "loss": 1.5732, "step": 2346 }, { "epoch": 0.2536474656868043, "grad_norm": 0.36444783210754395, "learning_rate": 8e-05, "loss": 1.693, "step": 2347 }, { "epoch": 0.2537555387441911, "grad_norm": 0.37828055024147034, "learning_rate": 8e-05, "loss": 1.7184, "step": 2348 }, { "epoch": 0.25386361180157785, "grad_norm": 0.4031600058078766, "learning_rate": 8e-05, "loss": 1.8711, "step": 2349 }, { "epoch": 0.25397168485896465, "grad_norm": 0.3750914931297302, "learning_rate": 8e-05, "loss": 1.636, "step": 2350 }, { "epoch": 0.25407975791635146, "grad_norm": 0.372081458568573, "learning_rate": 8e-05, "loss": 1.728, "step": 2351 }, { "epoch": 0.25418783097373826, "grad_norm": 0.37560608983039856, "learning_rate": 8e-05, "loss": 1.6529, "step": 2352 }, { "epoch": 0.25429590403112506, "grad_norm": 0.38047584891319275, "learning_rate": 8e-05, "loss": 1.6707, "step": 2353 }, { "epoch": 0.2544039770885118, "grad_norm": 0.4073527753353119, "learning_rate": 8e-05, "loss": 1.8785, "step": 2354 }, { "epoch": 0.2545120501458986, "grad_norm": 0.4195411801338196, "learning_rate": 8e-05, "loss": 1.8149, "step": 2355 }, { "epoch": 0.2546201232032854, "grad_norm": 0.40978819131851196, "learning_rate": 8e-05, "loss": 1.8837, "step": 2356 }, { "epoch": 0.2547281962606722, "grad_norm": 0.36176303029060364, "learning_rate": 8e-05, "loss": 1.7321, "step": 2357 }, { "epoch": 0.254836269318059, "grad_norm": 0.3815341591835022, "learning_rate": 8e-05, "loss": 1.646, "step": 2358 }, { "epoch": 0.2549443423754458, "grad_norm": 0.3497124910354614, "learning_rate": 8e-05, "loss": 1.6696, "step": 2359 }, { "epoch": 0.2550524154328326, "grad_norm": 0.3657241761684418, "learning_rate": 8e-05, "loss": 1.6053, "step": 2360 }, { "epoch": 0.2551604884902194, "grad_norm": 0.3724481463432312, "learning_rate": 8e-05, "loss": 1.684, "step": 2361 }, { "epoch": 0.2552685615476062, "grad_norm": 0.3785184919834137, "learning_rate": 8e-05, "loss": 1.6805, "step": 2362 }, { "epoch": 0.255376634604993, "grad_norm": 0.3680363893508911, "learning_rate": 8e-05, "loss": 1.7499, "step": 2363 }, { "epoch": 0.2554847076623798, "grad_norm": 0.3570915758609772, "learning_rate": 8e-05, "loss": 1.6243, "step": 2364 }, { "epoch": 0.25559278071976654, "grad_norm": 0.39442989230155945, "learning_rate": 8e-05, "loss": 1.805, "step": 2365 }, { "epoch": 0.25570085377715335, "grad_norm": 0.3628694415092468, "learning_rate": 8e-05, "loss": 1.717, "step": 2366 }, { "epoch": 0.25580892683454015, "grad_norm": 0.3751349151134491, "learning_rate": 8e-05, "loss": 1.6543, "step": 2367 }, { "epoch": 0.25591699989192696, "grad_norm": 0.36008697748184204, "learning_rate": 8e-05, "loss": 1.6463, "step": 2368 }, { "epoch": 0.25602507294931376, "grad_norm": 0.37018558382987976, "learning_rate": 8e-05, "loss": 1.6328, "step": 2369 }, { "epoch": 0.2561331460067005, "grad_norm": 0.34837254881858826, "learning_rate": 8e-05, "loss": 1.5436, "step": 2370 }, { "epoch": 0.2562412190640873, "grad_norm": 0.3591398596763611, "learning_rate": 8e-05, "loss": 1.5317, "step": 2371 }, { "epoch": 0.2563492921214741, "grad_norm": 0.34367167949676514, "learning_rate": 8e-05, "loss": 1.5222, "step": 2372 }, { "epoch": 0.2564573651788609, "grad_norm": 0.37893447279930115, "learning_rate": 8e-05, "loss": 1.8652, "step": 2373 }, { "epoch": 0.2565654382362477, "grad_norm": 0.3768373131752014, "learning_rate": 8e-05, "loss": 1.7244, "step": 2374 }, { "epoch": 0.25667351129363447, "grad_norm": 0.3967050313949585, "learning_rate": 8e-05, "loss": 1.9085, "step": 2375 }, { "epoch": 0.2567815843510213, "grad_norm": 0.3844134211540222, "learning_rate": 8e-05, "loss": 1.6798, "step": 2376 }, { "epoch": 0.2568896574084081, "grad_norm": 0.37405160069465637, "learning_rate": 8e-05, "loss": 1.6963, "step": 2377 }, { "epoch": 0.2569977304657949, "grad_norm": 0.3459620177745819, "learning_rate": 8e-05, "loss": 1.531, "step": 2378 }, { "epoch": 0.2571058035231817, "grad_norm": 0.37010252475738525, "learning_rate": 8e-05, "loss": 1.6831, "step": 2379 }, { "epoch": 0.2572138765805685, "grad_norm": 0.4051039516925812, "learning_rate": 8e-05, "loss": 1.7643, "step": 2380 }, { "epoch": 0.25732194963795524, "grad_norm": 0.34516793489456177, "learning_rate": 8e-05, "loss": 1.652, "step": 2381 }, { "epoch": 0.25743002269534204, "grad_norm": 0.3632252812385559, "learning_rate": 8e-05, "loss": 1.6696, "step": 2382 }, { "epoch": 0.25753809575272885, "grad_norm": 0.37083593010902405, "learning_rate": 8e-05, "loss": 1.6783, "step": 2383 }, { "epoch": 0.25764616881011565, "grad_norm": 0.36727359890937805, "learning_rate": 8e-05, "loss": 1.7419, "step": 2384 }, { "epoch": 0.25775424186750245, "grad_norm": 0.38211095333099365, "learning_rate": 8e-05, "loss": 1.6724, "step": 2385 }, { "epoch": 0.2578623149248892, "grad_norm": 0.4013848900794983, "learning_rate": 8e-05, "loss": 1.6704, "step": 2386 }, { "epoch": 0.257970387982276, "grad_norm": 0.3575046956539154, "learning_rate": 8e-05, "loss": 1.7418, "step": 2387 }, { "epoch": 0.2580784610396628, "grad_norm": 0.36692652106285095, "learning_rate": 8e-05, "loss": 1.6918, "step": 2388 }, { "epoch": 0.2581865340970496, "grad_norm": 0.37888476252555847, "learning_rate": 8e-05, "loss": 1.8369, "step": 2389 }, { "epoch": 0.2582946071544364, "grad_norm": 0.3732790946960449, "learning_rate": 8e-05, "loss": 1.6603, "step": 2390 }, { "epoch": 0.25840268021182317, "grad_norm": 0.38367220759391785, "learning_rate": 8e-05, "loss": 1.7893, "step": 2391 }, { "epoch": 0.25851075326920997, "grad_norm": 0.36935535073280334, "learning_rate": 8e-05, "loss": 1.6683, "step": 2392 }, { "epoch": 0.2586188263265968, "grad_norm": 0.3913491666316986, "learning_rate": 8e-05, "loss": 1.8542, "step": 2393 }, { "epoch": 0.2587268993839836, "grad_norm": 0.38293763995170593, "learning_rate": 8e-05, "loss": 1.765, "step": 2394 }, { "epoch": 0.2588349724413704, "grad_norm": 0.3737819790840149, "learning_rate": 8e-05, "loss": 1.8052, "step": 2395 }, { "epoch": 0.2589430454987572, "grad_norm": 0.37721824645996094, "learning_rate": 8e-05, "loss": 1.6376, "step": 2396 }, { "epoch": 0.25905111855614393, "grad_norm": 0.4264904260635376, "learning_rate": 8e-05, "loss": 1.8299, "step": 2397 }, { "epoch": 0.25915919161353074, "grad_norm": 0.3576003611087799, "learning_rate": 8e-05, "loss": 1.7774, "step": 2398 }, { "epoch": 0.25926726467091754, "grad_norm": 0.3670266568660736, "learning_rate": 8e-05, "loss": 1.8063, "step": 2399 }, { "epoch": 0.25937533772830434, "grad_norm": 0.4331955015659332, "learning_rate": 8e-05, "loss": 1.4671, "step": 2400 }, { "epoch": 0.25948341078569115, "grad_norm": 0.3563782870769501, "learning_rate": 8e-05, "loss": 1.6389, "step": 2401 }, { "epoch": 0.2595914838430779, "grad_norm": 0.3500118553638458, "learning_rate": 8e-05, "loss": 1.7297, "step": 2402 }, { "epoch": 0.2596995569004647, "grad_norm": 0.3432362377643585, "learning_rate": 8e-05, "loss": 1.4502, "step": 2403 }, { "epoch": 0.2598076299578515, "grad_norm": 0.39019864797592163, "learning_rate": 8e-05, "loss": 1.6847, "step": 2404 }, { "epoch": 0.2599157030152383, "grad_norm": 0.36085912585258484, "learning_rate": 8e-05, "loss": 1.8191, "step": 2405 }, { "epoch": 0.2600237760726251, "grad_norm": 0.3707723915576935, "learning_rate": 8e-05, "loss": 1.7399, "step": 2406 }, { "epoch": 0.26013184913001186, "grad_norm": 0.3534475266933441, "learning_rate": 8e-05, "loss": 1.5932, "step": 2407 }, { "epoch": 0.26023992218739866, "grad_norm": 0.37683358788490295, "learning_rate": 8e-05, "loss": 1.6872, "step": 2408 }, { "epoch": 0.26034799524478547, "grad_norm": 0.3793697953224182, "learning_rate": 8e-05, "loss": 1.6994, "step": 2409 }, { "epoch": 0.26045606830217227, "grad_norm": 0.35983163118362427, "learning_rate": 8e-05, "loss": 1.778, "step": 2410 }, { "epoch": 0.2605641413595591, "grad_norm": 0.388555109500885, "learning_rate": 8e-05, "loss": 1.7395, "step": 2411 }, { "epoch": 0.2606722144169459, "grad_norm": 0.39065611362457275, "learning_rate": 8e-05, "loss": 1.873, "step": 2412 }, { "epoch": 0.26078028747433263, "grad_norm": 0.45261526107788086, "learning_rate": 8e-05, "loss": 1.6543, "step": 2413 }, { "epoch": 0.26088836053171943, "grad_norm": 0.4009803235530853, "learning_rate": 8e-05, "loss": 1.9119, "step": 2414 }, { "epoch": 0.26099643358910624, "grad_norm": 0.3706305921077728, "learning_rate": 8e-05, "loss": 1.7213, "step": 2415 }, { "epoch": 0.26110450664649304, "grad_norm": 0.3484938442707062, "learning_rate": 8e-05, "loss": 1.6855, "step": 2416 }, { "epoch": 0.26121257970387984, "grad_norm": 0.4164220690727234, "learning_rate": 8e-05, "loss": 1.8719, "step": 2417 }, { "epoch": 0.2613206527612666, "grad_norm": 0.3573210835456848, "learning_rate": 8e-05, "loss": 1.5876, "step": 2418 }, { "epoch": 0.2614287258186534, "grad_norm": 0.38837939500808716, "learning_rate": 8e-05, "loss": 1.8216, "step": 2419 }, { "epoch": 0.2615367988760402, "grad_norm": 0.3543044626712799, "learning_rate": 8e-05, "loss": 1.6882, "step": 2420 }, { "epoch": 0.261644871933427, "grad_norm": 0.3907885253429413, "learning_rate": 8e-05, "loss": 1.8667, "step": 2421 }, { "epoch": 0.2617529449908138, "grad_norm": 0.37509965896606445, "learning_rate": 8e-05, "loss": 1.8062, "step": 2422 }, { "epoch": 0.2618610180482006, "grad_norm": 0.3805215060710907, "learning_rate": 8e-05, "loss": 1.7785, "step": 2423 }, { "epoch": 0.26196909110558736, "grad_norm": 0.3731456696987152, "learning_rate": 8e-05, "loss": 1.5353, "step": 2424 }, { "epoch": 0.26207716416297416, "grad_norm": 0.393075555562973, "learning_rate": 8e-05, "loss": 1.8361, "step": 2425 }, { "epoch": 0.26218523722036097, "grad_norm": 0.4163043200969696, "learning_rate": 8e-05, "loss": 1.8546, "step": 2426 }, { "epoch": 0.26229331027774777, "grad_norm": 0.35960590839385986, "learning_rate": 8e-05, "loss": 1.6065, "step": 2427 }, { "epoch": 0.2624013833351346, "grad_norm": 0.37370970845222473, "learning_rate": 8e-05, "loss": 1.7678, "step": 2428 }, { "epoch": 0.2625094563925213, "grad_norm": 0.4729326069355011, "learning_rate": 8e-05, "loss": 1.8061, "step": 2429 }, { "epoch": 0.2626175294499081, "grad_norm": 0.39908185601234436, "learning_rate": 8e-05, "loss": 1.9572, "step": 2430 }, { "epoch": 0.26272560250729493, "grad_norm": 0.4235093593597412, "learning_rate": 8e-05, "loss": 2.0039, "step": 2431 }, { "epoch": 0.26283367556468173, "grad_norm": 0.3778579533100128, "learning_rate": 8e-05, "loss": 1.8555, "step": 2432 }, { "epoch": 0.26294174862206854, "grad_norm": 0.3497962951660156, "learning_rate": 8e-05, "loss": 1.7224, "step": 2433 }, { "epoch": 0.2630498216794553, "grad_norm": 0.3692166805267334, "learning_rate": 8e-05, "loss": 1.7883, "step": 2434 }, { "epoch": 0.2631578947368421, "grad_norm": 0.39336279034614563, "learning_rate": 8e-05, "loss": 1.9107, "step": 2435 }, { "epoch": 0.2632659677942289, "grad_norm": 0.38543298840522766, "learning_rate": 8e-05, "loss": 1.73, "step": 2436 }, { "epoch": 0.2633740408516157, "grad_norm": 0.40469375252723694, "learning_rate": 8e-05, "loss": 1.6788, "step": 2437 }, { "epoch": 0.2634821139090025, "grad_norm": 0.3772241175174713, "learning_rate": 8e-05, "loss": 1.7007, "step": 2438 }, { "epoch": 0.2635901869663893, "grad_norm": 0.3549844026565552, "learning_rate": 8e-05, "loss": 1.5512, "step": 2439 }, { "epoch": 0.26369826002377605, "grad_norm": 0.4116078317165375, "learning_rate": 8e-05, "loss": 1.857, "step": 2440 }, { "epoch": 0.26380633308116286, "grad_norm": 0.3834722340106964, "learning_rate": 8e-05, "loss": 1.7679, "step": 2441 }, { "epoch": 0.26391440613854966, "grad_norm": 0.35018327832221985, "learning_rate": 8e-05, "loss": 1.706, "step": 2442 }, { "epoch": 0.26402247919593647, "grad_norm": 0.3909597098827362, "learning_rate": 8e-05, "loss": 1.8188, "step": 2443 }, { "epoch": 0.26413055225332327, "grad_norm": 0.3648270070552826, "learning_rate": 8e-05, "loss": 1.7368, "step": 2444 }, { "epoch": 0.26423862531071, "grad_norm": 0.3591890335083008, "learning_rate": 8e-05, "loss": 1.6948, "step": 2445 }, { "epoch": 0.2643466983680968, "grad_norm": 0.3624780476093292, "learning_rate": 8e-05, "loss": 1.5988, "step": 2446 }, { "epoch": 0.2644547714254836, "grad_norm": 0.33978351950645447, "learning_rate": 8e-05, "loss": 1.5544, "step": 2447 }, { "epoch": 0.26456284448287043, "grad_norm": 0.42829594016075134, "learning_rate": 8e-05, "loss": 1.6167, "step": 2448 }, { "epoch": 0.26467091754025723, "grad_norm": 0.35077425837516785, "learning_rate": 8e-05, "loss": 1.6844, "step": 2449 }, { "epoch": 0.264778990597644, "grad_norm": 0.41539913415908813, "learning_rate": 8e-05, "loss": 1.8585, "step": 2450 }, { "epoch": 0.2648870636550308, "grad_norm": 0.36336368322372437, "learning_rate": 8e-05, "loss": 1.7025, "step": 2451 }, { "epoch": 0.2649951367124176, "grad_norm": 0.35743921995162964, "learning_rate": 8e-05, "loss": 1.7554, "step": 2452 }, { "epoch": 0.2651032097698044, "grad_norm": 0.41128188371658325, "learning_rate": 8e-05, "loss": 1.7489, "step": 2453 }, { "epoch": 0.2652112828271912, "grad_norm": 0.38887327909469604, "learning_rate": 8e-05, "loss": 1.8813, "step": 2454 }, { "epoch": 0.265319355884578, "grad_norm": 0.3744156062602997, "learning_rate": 8e-05, "loss": 1.6251, "step": 2455 }, { "epoch": 0.26542742894196475, "grad_norm": 0.4440920948982239, "learning_rate": 8e-05, "loss": 1.9888, "step": 2456 }, { "epoch": 0.26553550199935155, "grad_norm": 0.3793190121650696, "learning_rate": 8e-05, "loss": 1.6845, "step": 2457 }, { "epoch": 0.26564357505673836, "grad_norm": 0.3661603331565857, "learning_rate": 8e-05, "loss": 1.6758, "step": 2458 }, { "epoch": 0.26575164811412516, "grad_norm": 0.36692652106285095, "learning_rate": 8e-05, "loss": 1.5304, "step": 2459 }, { "epoch": 0.26585972117151196, "grad_norm": 0.39514121413230896, "learning_rate": 8e-05, "loss": 1.7794, "step": 2460 }, { "epoch": 0.2659677942288987, "grad_norm": 0.3803490698337555, "learning_rate": 8e-05, "loss": 1.7086, "step": 2461 }, { "epoch": 0.2660758672862855, "grad_norm": 0.4119362533092499, "learning_rate": 8e-05, "loss": 1.9486, "step": 2462 }, { "epoch": 0.2661839403436723, "grad_norm": 0.3772425055503845, "learning_rate": 8e-05, "loss": 1.7165, "step": 2463 }, { "epoch": 0.2662920134010591, "grad_norm": 0.37792736291885376, "learning_rate": 8e-05, "loss": 1.7267, "step": 2464 }, { "epoch": 0.26640008645844593, "grad_norm": 0.3979353606700897, "learning_rate": 8e-05, "loss": 1.7175, "step": 2465 }, { "epoch": 0.2665081595158327, "grad_norm": 0.3579016625881195, "learning_rate": 8e-05, "loss": 1.7184, "step": 2466 }, { "epoch": 0.2666162325732195, "grad_norm": 0.39699140191078186, "learning_rate": 8e-05, "loss": 1.7969, "step": 2467 }, { "epoch": 0.2667243056306063, "grad_norm": 0.36525195837020874, "learning_rate": 8e-05, "loss": 1.7161, "step": 2468 }, { "epoch": 0.2668323786879931, "grad_norm": 0.369540810585022, "learning_rate": 8e-05, "loss": 1.6264, "step": 2469 }, { "epoch": 0.2669404517453799, "grad_norm": 0.37650567293167114, "learning_rate": 8e-05, "loss": 1.7501, "step": 2470 }, { "epoch": 0.2670485248027667, "grad_norm": 0.33726438879966736, "learning_rate": 8e-05, "loss": 1.6949, "step": 2471 }, { "epoch": 0.26715659786015344, "grad_norm": 0.38358521461486816, "learning_rate": 8e-05, "loss": 1.7437, "step": 2472 }, { "epoch": 0.26726467091754025, "grad_norm": 0.3594326376914978, "learning_rate": 8e-05, "loss": 1.6827, "step": 2473 }, { "epoch": 0.26737274397492705, "grad_norm": 0.40659692883491516, "learning_rate": 8e-05, "loss": 1.369, "step": 2474 }, { "epoch": 0.26748081703231386, "grad_norm": 0.3636024594306946, "learning_rate": 8e-05, "loss": 1.6558, "step": 2475 }, { "epoch": 0.26758889008970066, "grad_norm": 0.3700731098651886, "learning_rate": 8e-05, "loss": 1.663, "step": 2476 }, { "epoch": 0.2676969631470874, "grad_norm": 0.3898457884788513, "learning_rate": 8e-05, "loss": 1.8759, "step": 2477 }, { "epoch": 0.2678050362044742, "grad_norm": 0.3857495188713074, "learning_rate": 8e-05, "loss": 1.5405, "step": 2478 }, { "epoch": 0.267913109261861, "grad_norm": 0.36643165349960327, "learning_rate": 8e-05, "loss": 1.7793, "step": 2479 }, { "epoch": 0.2680211823192478, "grad_norm": 0.3721325993537903, "learning_rate": 8e-05, "loss": 1.6934, "step": 2480 }, { "epoch": 0.2681292553766346, "grad_norm": 0.36624324321746826, "learning_rate": 8e-05, "loss": 1.7586, "step": 2481 }, { "epoch": 0.26823732843402137, "grad_norm": 0.36084693670272827, "learning_rate": 8e-05, "loss": 1.8299, "step": 2482 }, { "epoch": 0.2683454014914082, "grad_norm": 0.3541159927845001, "learning_rate": 8e-05, "loss": 1.6488, "step": 2483 }, { "epoch": 0.268453474548795, "grad_norm": 0.4120561182498932, "learning_rate": 8e-05, "loss": 1.7216, "step": 2484 }, { "epoch": 0.2685615476061818, "grad_norm": 0.3740830421447754, "learning_rate": 8e-05, "loss": 1.7564, "step": 2485 }, { "epoch": 0.2686696206635686, "grad_norm": 0.3531447649002075, "learning_rate": 8e-05, "loss": 1.6796, "step": 2486 }, { "epoch": 0.2687776937209554, "grad_norm": 0.3760266602039337, "learning_rate": 8e-05, "loss": 1.8165, "step": 2487 }, { "epoch": 0.26888576677834214, "grad_norm": 0.4254160225391388, "learning_rate": 8e-05, "loss": 1.4274, "step": 2488 }, { "epoch": 0.26899383983572894, "grad_norm": 0.34750789403915405, "learning_rate": 8e-05, "loss": 1.4991, "step": 2489 }, { "epoch": 0.26910191289311575, "grad_norm": 0.4012928605079651, "learning_rate": 8e-05, "loss": 1.8025, "step": 2490 }, { "epoch": 0.26920998595050255, "grad_norm": 0.3861410915851593, "learning_rate": 8e-05, "loss": 1.6657, "step": 2491 }, { "epoch": 0.26931805900788935, "grad_norm": 0.3659304678440094, "learning_rate": 8e-05, "loss": 1.7837, "step": 2492 }, { "epoch": 0.2694261320652761, "grad_norm": 0.42097416520118713, "learning_rate": 8e-05, "loss": 1.8566, "step": 2493 }, { "epoch": 0.2695342051226629, "grad_norm": 0.36721569299697876, "learning_rate": 8e-05, "loss": 1.7708, "step": 2494 }, { "epoch": 0.2696422781800497, "grad_norm": 0.40271109342575073, "learning_rate": 8e-05, "loss": 1.6643, "step": 2495 }, { "epoch": 0.2697503512374365, "grad_norm": 0.37576308846473694, "learning_rate": 8e-05, "loss": 1.8026, "step": 2496 }, { "epoch": 0.2698584242948233, "grad_norm": 0.3525030314922333, "learning_rate": 8e-05, "loss": 1.7487, "step": 2497 }, { "epoch": 0.26996649735221007, "grad_norm": 0.41718024015426636, "learning_rate": 8e-05, "loss": 1.8849, "step": 2498 }, { "epoch": 0.27007457040959687, "grad_norm": 0.382348895072937, "learning_rate": 8e-05, "loss": 1.7286, "step": 2499 }, { "epoch": 0.2701826434669837, "grad_norm": 0.4037102460861206, "learning_rate": 8e-05, "loss": 1.7359, "step": 2500 }, { "epoch": 0.2702907165243705, "grad_norm": 0.37202468514442444, "learning_rate": 8e-05, "loss": 1.7424, "step": 2501 }, { "epoch": 0.2703987895817573, "grad_norm": 0.3814954161643982, "learning_rate": 8e-05, "loss": 1.7672, "step": 2502 }, { "epoch": 0.2705068626391441, "grad_norm": 0.37354743480682373, "learning_rate": 8e-05, "loss": 1.6408, "step": 2503 }, { "epoch": 0.27061493569653083, "grad_norm": 0.3903687298297882, "learning_rate": 8e-05, "loss": 1.7209, "step": 2504 }, { "epoch": 0.27072300875391764, "grad_norm": 0.4233676791191101, "learning_rate": 8e-05, "loss": 1.8452, "step": 2505 }, { "epoch": 0.27083108181130444, "grad_norm": 0.3956805169582367, "learning_rate": 8e-05, "loss": 1.7678, "step": 2506 }, { "epoch": 0.27093915486869125, "grad_norm": 0.3765985369682312, "learning_rate": 8e-05, "loss": 1.6997, "step": 2507 }, { "epoch": 0.27104722792607805, "grad_norm": 0.3490865230560303, "learning_rate": 8e-05, "loss": 1.5437, "step": 2508 }, { "epoch": 0.2711553009834648, "grad_norm": 0.3582180440425873, "learning_rate": 8e-05, "loss": 1.6532, "step": 2509 }, { "epoch": 0.2712633740408516, "grad_norm": 0.37423941493034363, "learning_rate": 8e-05, "loss": 1.6049, "step": 2510 }, { "epoch": 0.2713714470982384, "grad_norm": 0.39894378185272217, "learning_rate": 8e-05, "loss": 1.6756, "step": 2511 }, { "epoch": 0.2714795201556252, "grad_norm": 0.3725018799304962, "learning_rate": 8e-05, "loss": 1.7469, "step": 2512 }, { "epoch": 0.271587593213012, "grad_norm": 0.3482353091239929, "learning_rate": 8e-05, "loss": 1.5849, "step": 2513 }, { "epoch": 0.2716956662703988, "grad_norm": 0.3945613205432892, "learning_rate": 8e-05, "loss": 1.8173, "step": 2514 }, { "epoch": 0.27180373932778557, "grad_norm": 0.3775917589664459, "learning_rate": 8e-05, "loss": 1.8485, "step": 2515 }, { "epoch": 0.27191181238517237, "grad_norm": 0.36745280027389526, "learning_rate": 8e-05, "loss": 1.682, "step": 2516 }, { "epoch": 0.2720198854425592, "grad_norm": 0.36604225635528564, "learning_rate": 8e-05, "loss": 1.7279, "step": 2517 }, { "epoch": 0.272127958499946, "grad_norm": 0.4445498585700989, "learning_rate": 8e-05, "loss": 2.0939, "step": 2518 }, { "epoch": 0.2722360315573328, "grad_norm": 0.3474128842353821, "learning_rate": 8e-05, "loss": 1.613, "step": 2519 }, { "epoch": 0.27234410461471953, "grad_norm": 0.35501158237457275, "learning_rate": 8e-05, "loss": 1.5669, "step": 2520 }, { "epoch": 0.27245217767210633, "grad_norm": 0.38934776186943054, "learning_rate": 8e-05, "loss": 1.8804, "step": 2521 }, { "epoch": 0.27256025072949314, "grad_norm": 0.37549129128456116, "learning_rate": 8e-05, "loss": 1.8113, "step": 2522 }, { "epoch": 0.27266832378687994, "grad_norm": 0.34787410497665405, "learning_rate": 8e-05, "loss": 1.5699, "step": 2523 }, { "epoch": 0.27277639684426674, "grad_norm": 0.3693847358226776, "learning_rate": 8e-05, "loss": 1.6059, "step": 2524 }, { "epoch": 0.2728844699016535, "grad_norm": 0.380143404006958, "learning_rate": 8e-05, "loss": 1.7746, "step": 2525 }, { "epoch": 0.2729925429590403, "grad_norm": 0.3778119385242462, "learning_rate": 8e-05, "loss": 1.7387, "step": 2526 }, { "epoch": 0.2731006160164271, "grad_norm": 0.3620222210884094, "learning_rate": 8e-05, "loss": 1.6545, "step": 2527 }, { "epoch": 0.2732086890738139, "grad_norm": 0.36658886075019836, "learning_rate": 8e-05, "loss": 1.7734, "step": 2528 }, { "epoch": 0.2733167621312007, "grad_norm": 0.413739949464798, "learning_rate": 8e-05, "loss": 1.7863, "step": 2529 }, { "epoch": 0.2734248351885875, "grad_norm": 0.37141475081443787, "learning_rate": 8e-05, "loss": 1.7975, "step": 2530 }, { "epoch": 0.27353290824597426, "grad_norm": 0.3921930491924286, "learning_rate": 8e-05, "loss": 1.7961, "step": 2531 }, { "epoch": 0.27364098130336106, "grad_norm": 0.3504860997200012, "learning_rate": 8e-05, "loss": 1.7217, "step": 2532 }, { "epoch": 0.27374905436074787, "grad_norm": 0.374895840883255, "learning_rate": 8e-05, "loss": 1.8226, "step": 2533 }, { "epoch": 0.27385712741813467, "grad_norm": 0.392419695854187, "learning_rate": 8e-05, "loss": 1.7938, "step": 2534 }, { "epoch": 0.2739652004755215, "grad_norm": 0.35362112522125244, "learning_rate": 8e-05, "loss": 1.7076, "step": 2535 }, { "epoch": 0.2740732735329082, "grad_norm": 0.3471091091632843, "learning_rate": 8e-05, "loss": 1.5148, "step": 2536 }, { "epoch": 0.274181346590295, "grad_norm": 0.3957866430282593, "learning_rate": 8e-05, "loss": 1.8398, "step": 2537 }, { "epoch": 0.27428941964768183, "grad_norm": 0.3853392004966736, "learning_rate": 8e-05, "loss": 1.7186, "step": 2538 }, { "epoch": 0.27439749270506864, "grad_norm": 0.3762071430683136, "learning_rate": 8e-05, "loss": 1.6908, "step": 2539 }, { "epoch": 0.27450556576245544, "grad_norm": 0.3879365026950836, "learning_rate": 8e-05, "loss": 1.5545, "step": 2540 }, { "epoch": 0.2746136388198422, "grad_norm": 0.3753284811973572, "learning_rate": 8e-05, "loss": 1.6732, "step": 2541 }, { "epoch": 0.274721711877229, "grad_norm": 0.4130113422870636, "learning_rate": 8e-05, "loss": 1.8967, "step": 2542 }, { "epoch": 0.2748297849346158, "grad_norm": 0.4417731761932373, "learning_rate": 8e-05, "loss": 1.8916, "step": 2543 }, { "epoch": 0.2749378579920026, "grad_norm": 0.35213807225227356, "learning_rate": 8e-05, "loss": 1.7227, "step": 2544 }, { "epoch": 0.2750459310493894, "grad_norm": 0.38566669821739197, "learning_rate": 8e-05, "loss": 1.8435, "step": 2545 }, { "epoch": 0.2751540041067762, "grad_norm": 0.36589381098747253, "learning_rate": 8e-05, "loss": 1.6623, "step": 2546 }, { "epoch": 0.27526207716416295, "grad_norm": 0.3881056308746338, "learning_rate": 8e-05, "loss": 1.7444, "step": 2547 }, { "epoch": 0.27537015022154976, "grad_norm": 0.35593181848526, "learning_rate": 8e-05, "loss": 1.6498, "step": 2548 }, { "epoch": 0.27547822327893656, "grad_norm": 0.38285768032073975, "learning_rate": 8e-05, "loss": 1.8168, "step": 2549 }, { "epoch": 0.27558629633632337, "grad_norm": 0.34929969906806946, "learning_rate": 8e-05, "loss": 1.6237, "step": 2550 }, { "epoch": 0.27569436939371017, "grad_norm": 0.38450849056243896, "learning_rate": 8e-05, "loss": 1.7313, "step": 2551 }, { "epoch": 0.2758024424510969, "grad_norm": 0.40309420228004456, "learning_rate": 8e-05, "loss": 1.877, "step": 2552 }, { "epoch": 0.2759105155084837, "grad_norm": 0.37091633677482605, "learning_rate": 8e-05, "loss": 1.7201, "step": 2553 }, { "epoch": 0.2760185885658705, "grad_norm": 0.3870861232280731, "learning_rate": 8e-05, "loss": 1.7462, "step": 2554 }, { "epoch": 0.27612666162325733, "grad_norm": 0.4026245176792145, "learning_rate": 8e-05, "loss": 1.848, "step": 2555 }, { "epoch": 0.27623473468064413, "grad_norm": 0.43537789583206177, "learning_rate": 8e-05, "loss": 1.826, "step": 2556 }, { "epoch": 0.2763428077380309, "grad_norm": 0.3637314736843109, "learning_rate": 8e-05, "loss": 1.8177, "step": 2557 }, { "epoch": 0.2764508807954177, "grad_norm": 0.3755403459072113, "learning_rate": 8e-05, "loss": 1.5909, "step": 2558 }, { "epoch": 0.2765589538528045, "grad_norm": 0.4269295632839203, "learning_rate": 8e-05, "loss": 1.86, "step": 2559 }, { "epoch": 0.2766670269101913, "grad_norm": 0.4731540381908417, "learning_rate": 8e-05, "loss": 1.8046, "step": 2560 }, { "epoch": 0.2767750999675781, "grad_norm": 0.43560025095939636, "learning_rate": 8e-05, "loss": 1.8978, "step": 2561 }, { "epoch": 0.2768831730249649, "grad_norm": 0.362594872713089, "learning_rate": 8e-05, "loss": 1.6746, "step": 2562 }, { "epoch": 0.27699124608235165, "grad_norm": 0.3803568482398987, "learning_rate": 8e-05, "loss": 1.6798, "step": 2563 }, { "epoch": 0.27709931913973845, "grad_norm": 0.42912617325782776, "learning_rate": 8e-05, "loss": 1.7271, "step": 2564 }, { "epoch": 0.27720739219712526, "grad_norm": 0.382752388715744, "learning_rate": 8e-05, "loss": 1.6952, "step": 2565 }, { "epoch": 0.27731546525451206, "grad_norm": 0.3976055979728699, "learning_rate": 8e-05, "loss": 1.8627, "step": 2566 }, { "epoch": 0.27742353831189887, "grad_norm": 0.43283385038375854, "learning_rate": 8e-05, "loss": 1.9287, "step": 2567 }, { "epoch": 0.2775316113692856, "grad_norm": 0.406622976064682, "learning_rate": 8e-05, "loss": 1.6796, "step": 2568 }, { "epoch": 0.2776396844266724, "grad_norm": 0.43080392479896545, "learning_rate": 8e-05, "loss": 1.8439, "step": 2569 }, { "epoch": 0.2777477574840592, "grad_norm": 0.3766738474369049, "learning_rate": 8e-05, "loss": 1.7337, "step": 2570 }, { "epoch": 0.277855830541446, "grad_norm": 0.3888819217681885, "learning_rate": 8e-05, "loss": 1.7595, "step": 2571 }, { "epoch": 0.27796390359883283, "grad_norm": 0.3656492531299591, "learning_rate": 8e-05, "loss": 1.4745, "step": 2572 }, { "epoch": 0.2780719766562196, "grad_norm": 0.37483128905296326, "learning_rate": 8e-05, "loss": 1.7019, "step": 2573 }, { "epoch": 0.2781800497136064, "grad_norm": 0.3898029029369354, "learning_rate": 8e-05, "loss": 1.5681, "step": 2574 }, { "epoch": 0.2782881227709932, "grad_norm": 0.39038920402526855, "learning_rate": 8e-05, "loss": 1.8807, "step": 2575 }, { "epoch": 0.27839619582838, "grad_norm": 0.36786362528800964, "learning_rate": 8e-05, "loss": 1.7198, "step": 2576 }, { "epoch": 0.2785042688857668, "grad_norm": 0.37234652042388916, "learning_rate": 8e-05, "loss": 1.633, "step": 2577 }, { "epoch": 0.2786123419431536, "grad_norm": 0.3831823468208313, "learning_rate": 8e-05, "loss": 1.7557, "step": 2578 }, { "epoch": 0.27872041500054034, "grad_norm": 0.4305756986141205, "learning_rate": 8e-05, "loss": 1.8839, "step": 2579 }, { "epoch": 0.27882848805792715, "grad_norm": 0.3725205063819885, "learning_rate": 8e-05, "loss": 1.7784, "step": 2580 }, { "epoch": 0.27893656111531395, "grad_norm": 0.38674360513687134, "learning_rate": 8e-05, "loss": 1.7909, "step": 2581 }, { "epoch": 0.27904463417270076, "grad_norm": 0.4043315052986145, "learning_rate": 8e-05, "loss": 1.9964, "step": 2582 }, { "epoch": 0.27915270723008756, "grad_norm": 0.37745967507362366, "learning_rate": 8e-05, "loss": 1.7616, "step": 2583 }, { "epoch": 0.2792607802874743, "grad_norm": 0.38301882147789, "learning_rate": 8e-05, "loss": 1.8636, "step": 2584 }, { "epoch": 0.2793688533448611, "grad_norm": 0.38333404064178467, "learning_rate": 8e-05, "loss": 1.8089, "step": 2585 }, { "epoch": 0.2794769264022479, "grad_norm": 0.3695251941680908, "learning_rate": 8e-05, "loss": 1.6163, "step": 2586 }, { "epoch": 0.2795849994596347, "grad_norm": 0.3937545418739319, "learning_rate": 8e-05, "loss": 1.7227, "step": 2587 }, { "epoch": 0.2796930725170215, "grad_norm": 0.3846130967140198, "learning_rate": 8e-05, "loss": 1.7385, "step": 2588 }, { "epoch": 0.2798011455744083, "grad_norm": 0.34087884426116943, "learning_rate": 8e-05, "loss": 1.5742, "step": 2589 }, { "epoch": 0.2799092186317951, "grad_norm": 0.3758355379104614, "learning_rate": 8e-05, "loss": 1.7123, "step": 2590 }, { "epoch": 0.2800172916891819, "grad_norm": 0.3709929585456848, "learning_rate": 8e-05, "loss": 1.6953, "step": 2591 }, { "epoch": 0.2801253647465687, "grad_norm": 0.39020630717277527, "learning_rate": 8e-05, "loss": 1.7082, "step": 2592 }, { "epoch": 0.2802334378039555, "grad_norm": 0.3844773769378662, "learning_rate": 8e-05, "loss": 1.7821, "step": 2593 }, { "epoch": 0.2803415108613423, "grad_norm": 0.3910813629627228, "learning_rate": 8e-05, "loss": 1.8336, "step": 2594 }, { "epoch": 0.28044958391872904, "grad_norm": 0.3727092742919922, "learning_rate": 8e-05, "loss": 1.609, "step": 2595 }, { "epoch": 0.28055765697611584, "grad_norm": 0.3810289204120636, "learning_rate": 8e-05, "loss": 1.7389, "step": 2596 }, { "epoch": 0.28066573003350265, "grad_norm": 0.4203589856624603, "learning_rate": 8e-05, "loss": 1.8516, "step": 2597 }, { "epoch": 0.28077380309088945, "grad_norm": 0.36351585388183594, "learning_rate": 8e-05, "loss": 1.7073, "step": 2598 }, { "epoch": 0.28088187614827625, "grad_norm": 0.373172402381897, "learning_rate": 8e-05, "loss": 1.6788, "step": 2599 }, { "epoch": 0.280989949205663, "grad_norm": 0.35284096002578735, "learning_rate": 8e-05, "loss": 1.6921, "step": 2600 }, { "epoch": 0.2810980222630498, "grad_norm": 0.40052998065948486, "learning_rate": 8e-05, "loss": 1.8309, "step": 2601 }, { "epoch": 0.2812060953204366, "grad_norm": 0.38429924845695496, "learning_rate": 8e-05, "loss": 1.6739, "step": 2602 }, { "epoch": 0.2813141683778234, "grad_norm": 0.37491101026535034, "learning_rate": 8e-05, "loss": 1.8476, "step": 2603 }, { "epoch": 0.2814222414352102, "grad_norm": 0.398975133895874, "learning_rate": 8e-05, "loss": 1.803, "step": 2604 }, { "epoch": 0.281530314492597, "grad_norm": 0.3995000422000885, "learning_rate": 8e-05, "loss": 1.8905, "step": 2605 }, { "epoch": 0.28163838754998377, "grad_norm": 0.37232160568237305, "learning_rate": 8e-05, "loss": 1.5507, "step": 2606 }, { "epoch": 0.2817464606073706, "grad_norm": 0.3866870701313019, "learning_rate": 8e-05, "loss": 1.8756, "step": 2607 }, { "epoch": 0.2818545336647574, "grad_norm": 0.371979683637619, "learning_rate": 8e-05, "loss": 1.8679, "step": 2608 }, { "epoch": 0.2819626067221442, "grad_norm": 0.3651829659938812, "learning_rate": 8e-05, "loss": 1.6575, "step": 2609 }, { "epoch": 0.282070679779531, "grad_norm": 0.3690573275089264, "learning_rate": 8e-05, "loss": 1.6165, "step": 2610 }, { "epoch": 0.28217875283691773, "grad_norm": 0.3713589906692505, "learning_rate": 8e-05, "loss": 1.5919, "step": 2611 }, { "epoch": 0.28228682589430454, "grad_norm": 0.3814350962638855, "learning_rate": 8e-05, "loss": 1.708, "step": 2612 }, { "epoch": 0.28239489895169134, "grad_norm": 0.40340250730514526, "learning_rate": 8e-05, "loss": 1.862, "step": 2613 }, { "epoch": 0.28250297200907815, "grad_norm": 0.3941473364830017, "learning_rate": 8e-05, "loss": 1.9048, "step": 2614 }, { "epoch": 0.28261104506646495, "grad_norm": 0.41903766989707947, "learning_rate": 8e-05, "loss": 1.8632, "step": 2615 }, { "epoch": 0.2827191181238517, "grad_norm": 0.3816618323326111, "learning_rate": 8e-05, "loss": 1.7798, "step": 2616 }, { "epoch": 0.2828271911812385, "grad_norm": 0.3765718936920166, "learning_rate": 8e-05, "loss": 1.5974, "step": 2617 }, { "epoch": 0.2829352642386253, "grad_norm": 0.4068194329738617, "learning_rate": 8e-05, "loss": 1.6436, "step": 2618 }, { "epoch": 0.2830433372960121, "grad_norm": 0.3559759855270386, "learning_rate": 8e-05, "loss": 1.6816, "step": 2619 }, { "epoch": 0.2831514103533989, "grad_norm": 0.36500853300094604, "learning_rate": 8e-05, "loss": 1.6264, "step": 2620 }, { "epoch": 0.2832594834107857, "grad_norm": 0.3958582580089569, "learning_rate": 8e-05, "loss": 1.5652, "step": 2621 }, { "epoch": 0.28336755646817247, "grad_norm": 0.4033443331718445, "learning_rate": 8e-05, "loss": 1.7828, "step": 2622 }, { "epoch": 0.28347562952555927, "grad_norm": 0.3446609675884247, "learning_rate": 8e-05, "loss": 1.5903, "step": 2623 }, { "epoch": 0.2835837025829461, "grad_norm": 0.40247175097465515, "learning_rate": 8e-05, "loss": 1.8313, "step": 2624 }, { "epoch": 0.2836917756403329, "grad_norm": 0.4228101372718811, "learning_rate": 8e-05, "loss": 1.8268, "step": 2625 }, { "epoch": 0.2837998486977197, "grad_norm": 0.4078630208969116, "learning_rate": 8e-05, "loss": 1.8554, "step": 2626 }, { "epoch": 0.28390792175510643, "grad_norm": 0.36746469140052795, "learning_rate": 8e-05, "loss": 1.7931, "step": 2627 }, { "epoch": 0.28401599481249323, "grad_norm": 0.3927038609981537, "learning_rate": 8e-05, "loss": 1.817, "step": 2628 }, { "epoch": 0.28412406786988004, "grad_norm": 0.37484753131866455, "learning_rate": 8e-05, "loss": 1.6864, "step": 2629 }, { "epoch": 0.28423214092726684, "grad_norm": 0.3721831440925598, "learning_rate": 8e-05, "loss": 1.6445, "step": 2630 }, { "epoch": 0.28434021398465364, "grad_norm": 0.3869171142578125, "learning_rate": 8e-05, "loss": 1.7173, "step": 2631 }, { "epoch": 0.2844482870420404, "grad_norm": 0.37088584899902344, "learning_rate": 8e-05, "loss": 1.585, "step": 2632 }, { "epoch": 0.2845563600994272, "grad_norm": 0.3893742859363556, "learning_rate": 8e-05, "loss": 1.767, "step": 2633 }, { "epoch": 0.284664433156814, "grad_norm": 0.3971605896949768, "learning_rate": 8e-05, "loss": 1.7925, "step": 2634 }, { "epoch": 0.2847725062142008, "grad_norm": 0.3455650806427002, "learning_rate": 8e-05, "loss": 1.7777, "step": 2635 }, { "epoch": 0.2848805792715876, "grad_norm": 0.383194237947464, "learning_rate": 8e-05, "loss": 1.7238, "step": 2636 }, { "epoch": 0.2849886523289744, "grad_norm": 0.39578551054000854, "learning_rate": 8e-05, "loss": 1.7229, "step": 2637 }, { "epoch": 0.28509672538636116, "grad_norm": 0.3673846423625946, "learning_rate": 8e-05, "loss": 1.6656, "step": 2638 }, { "epoch": 0.28520479844374796, "grad_norm": 0.40245911478996277, "learning_rate": 8e-05, "loss": 1.9308, "step": 2639 }, { "epoch": 0.28531287150113477, "grad_norm": 0.35533931851387024, "learning_rate": 8e-05, "loss": 1.5659, "step": 2640 }, { "epoch": 0.28542094455852157, "grad_norm": 0.36477547883987427, "learning_rate": 8e-05, "loss": 1.5876, "step": 2641 }, { "epoch": 0.2855290176159084, "grad_norm": 0.3914869427680969, "learning_rate": 8e-05, "loss": 1.7271, "step": 2642 }, { "epoch": 0.2856370906732951, "grad_norm": 0.3821012079715729, "learning_rate": 8e-05, "loss": 1.8567, "step": 2643 }, { "epoch": 0.28574516373068193, "grad_norm": 0.4113089144229889, "learning_rate": 8e-05, "loss": 2.0372, "step": 2644 }, { "epoch": 0.28585323678806873, "grad_norm": 0.3847944438457489, "learning_rate": 8e-05, "loss": 1.6981, "step": 2645 }, { "epoch": 0.28596130984545554, "grad_norm": 0.386808842420578, "learning_rate": 8e-05, "loss": 1.7791, "step": 2646 }, { "epoch": 0.28606938290284234, "grad_norm": 0.39247071743011475, "learning_rate": 8e-05, "loss": 1.7785, "step": 2647 }, { "epoch": 0.2861774559602291, "grad_norm": 0.3872281014919281, "learning_rate": 8e-05, "loss": 1.7778, "step": 2648 }, { "epoch": 0.2862855290176159, "grad_norm": 0.4142041504383087, "learning_rate": 8e-05, "loss": 1.801, "step": 2649 }, { "epoch": 0.2863936020750027, "grad_norm": 0.38411226868629456, "learning_rate": 8e-05, "loss": 1.649, "step": 2650 }, { "epoch": 0.2865016751323895, "grad_norm": 0.3829296827316284, "learning_rate": 8e-05, "loss": 1.7458, "step": 2651 }, { "epoch": 0.2866097481897763, "grad_norm": 0.37057220935821533, "learning_rate": 8e-05, "loss": 1.8394, "step": 2652 }, { "epoch": 0.2867178212471631, "grad_norm": 0.37490999698638916, "learning_rate": 8e-05, "loss": 1.6782, "step": 2653 }, { "epoch": 0.28682589430454986, "grad_norm": 0.3798205554485321, "learning_rate": 8e-05, "loss": 1.7162, "step": 2654 }, { "epoch": 0.28693396736193666, "grad_norm": 0.42090654373168945, "learning_rate": 8e-05, "loss": 1.576, "step": 2655 }, { "epoch": 0.28704204041932346, "grad_norm": 0.363699734210968, "learning_rate": 8e-05, "loss": 1.7029, "step": 2656 }, { "epoch": 0.28715011347671027, "grad_norm": 0.36247774958610535, "learning_rate": 8e-05, "loss": 1.6486, "step": 2657 }, { "epoch": 0.28725818653409707, "grad_norm": 0.4152909815311432, "learning_rate": 8e-05, "loss": 1.805, "step": 2658 }, { "epoch": 0.2873662595914838, "grad_norm": 0.39140433073043823, "learning_rate": 8e-05, "loss": 1.9045, "step": 2659 }, { "epoch": 0.2874743326488706, "grad_norm": 0.4135834872722626, "learning_rate": 8e-05, "loss": 1.9079, "step": 2660 }, { "epoch": 0.2875824057062574, "grad_norm": 0.43586188554763794, "learning_rate": 8e-05, "loss": 2.0383, "step": 2661 }, { "epoch": 0.28769047876364423, "grad_norm": 0.3908523619174957, "learning_rate": 8e-05, "loss": 1.7885, "step": 2662 }, { "epoch": 0.28779855182103103, "grad_norm": 0.4065014123916626, "learning_rate": 8e-05, "loss": 1.6589, "step": 2663 }, { "epoch": 0.2879066248784178, "grad_norm": 0.3792342245578766, "learning_rate": 8e-05, "loss": 1.8533, "step": 2664 }, { "epoch": 0.2880146979358046, "grad_norm": 0.3709142506122589, "learning_rate": 8e-05, "loss": 1.8187, "step": 2665 }, { "epoch": 0.2881227709931914, "grad_norm": 0.37438422441482544, "learning_rate": 8e-05, "loss": 1.6751, "step": 2666 }, { "epoch": 0.2882308440505782, "grad_norm": 0.3955494165420532, "learning_rate": 8e-05, "loss": 1.6588, "step": 2667 }, { "epoch": 0.288338917107965, "grad_norm": 0.3724851906299591, "learning_rate": 8e-05, "loss": 1.6725, "step": 2668 }, { "epoch": 0.2884469901653518, "grad_norm": 0.36863845586776733, "learning_rate": 8e-05, "loss": 1.6188, "step": 2669 }, { "epoch": 0.28855506322273855, "grad_norm": 0.40622618794441223, "learning_rate": 8e-05, "loss": 1.9388, "step": 2670 }, { "epoch": 0.28866313628012535, "grad_norm": 0.4573462903499603, "learning_rate": 8e-05, "loss": 1.9513, "step": 2671 }, { "epoch": 0.28877120933751216, "grad_norm": 0.37362560629844666, "learning_rate": 8e-05, "loss": 1.7882, "step": 2672 }, { "epoch": 0.28887928239489896, "grad_norm": 0.4047364890575409, "learning_rate": 8e-05, "loss": 1.9614, "step": 2673 }, { "epoch": 0.28898735545228577, "grad_norm": 0.37733224034309387, "learning_rate": 8e-05, "loss": 1.7854, "step": 2674 }, { "epoch": 0.2890954285096725, "grad_norm": 0.3852332830429077, "learning_rate": 8e-05, "loss": 1.6166, "step": 2675 }, { "epoch": 0.2892035015670593, "grad_norm": 0.39140352606773376, "learning_rate": 8e-05, "loss": 1.7995, "step": 2676 }, { "epoch": 0.2893115746244461, "grad_norm": 0.4005804657936096, "learning_rate": 8e-05, "loss": 1.8852, "step": 2677 }, { "epoch": 0.2894196476818329, "grad_norm": 0.3712862730026245, "learning_rate": 8e-05, "loss": 1.4719, "step": 2678 }, { "epoch": 0.28952772073921973, "grad_norm": 0.3938542604446411, "learning_rate": 8e-05, "loss": 1.6248, "step": 2679 }, { "epoch": 0.28963579379660653, "grad_norm": 0.3647788465023041, "learning_rate": 8e-05, "loss": 1.7077, "step": 2680 }, { "epoch": 0.2897438668539933, "grad_norm": 0.3877812325954437, "learning_rate": 8e-05, "loss": 1.6476, "step": 2681 }, { "epoch": 0.2898519399113801, "grad_norm": 0.35888177156448364, "learning_rate": 8e-05, "loss": 1.7229, "step": 2682 }, { "epoch": 0.2899600129687669, "grad_norm": 0.3767542243003845, "learning_rate": 8e-05, "loss": 1.7956, "step": 2683 }, { "epoch": 0.2900680860261537, "grad_norm": 0.3777533173561096, "learning_rate": 8e-05, "loss": 1.8302, "step": 2684 }, { "epoch": 0.2901761590835405, "grad_norm": 0.36561882495880127, "learning_rate": 8e-05, "loss": 1.8739, "step": 2685 }, { "epoch": 0.29028423214092725, "grad_norm": 0.39992254972457886, "learning_rate": 8e-05, "loss": 1.7131, "step": 2686 }, { "epoch": 0.29039230519831405, "grad_norm": 0.3876323699951172, "learning_rate": 8e-05, "loss": 1.5777, "step": 2687 }, { "epoch": 0.29050037825570085, "grad_norm": 0.3957829475402832, "learning_rate": 8e-05, "loss": 1.7931, "step": 2688 }, { "epoch": 0.29060845131308766, "grad_norm": 0.40311309695243835, "learning_rate": 8e-05, "loss": 1.6922, "step": 2689 }, { "epoch": 0.29071652437047446, "grad_norm": 0.40472134947776794, "learning_rate": 8e-05, "loss": 1.7055, "step": 2690 }, { "epoch": 0.2908245974278612, "grad_norm": 0.3906451463699341, "learning_rate": 8e-05, "loss": 1.7771, "step": 2691 }, { "epoch": 0.290932670485248, "grad_norm": 0.4069863259792328, "learning_rate": 8e-05, "loss": 1.8099, "step": 2692 }, { "epoch": 0.2910407435426348, "grad_norm": 0.3760221004486084, "learning_rate": 8e-05, "loss": 1.7386, "step": 2693 }, { "epoch": 0.2911488166000216, "grad_norm": 0.43671557307243347, "learning_rate": 8e-05, "loss": 1.8761, "step": 2694 }, { "epoch": 0.2912568896574084, "grad_norm": 0.38148486614227295, "learning_rate": 8e-05, "loss": 1.8413, "step": 2695 }, { "epoch": 0.29136496271479523, "grad_norm": 0.3758547604084015, "learning_rate": 8e-05, "loss": 1.7192, "step": 2696 }, { "epoch": 0.291473035772182, "grad_norm": 0.4103824496269226, "learning_rate": 8e-05, "loss": 1.8662, "step": 2697 }, { "epoch": 0.2915811088295688, "grad_norm": 0.34544193744659424, "learning_rate": 8e-05, "loss": 1.601, "step": 2698 }, { "epoch": 0.2916891818869556, "grad_norm": 0.39112764596939087, "learning_rate": 8e-05, "loss": 1.8879, "step": 2699 }, { "epoch": 0.2917972549443424, "grad_norm": 0.40787893533706665, "learning_rate": 8e-05, "loss": 1.7701, "step": 2700 }, { "epoch": 0.2919053280017292, "grad_norm": 0.3879365622997284, "learning_rate": 8e-05, "loss": 1.7712, "step": 2701 }, { "epoch": 0.29201340105911594, "grad_norm": 0.3981607258319855, "learning_rate": 8e-05, "loss": 1.8479, "step": 2702 }, { "epoch": 0.29212147411650274, "grad_norm": 0.39715006947517395, "learning_rate": 8e-05, "loss": 1.6604, "step": 2703 }, { "epoch": 0.29222954717388955, "grad_norm": 0.3840593099594116, "learning_rate": 8e-05, "loss": 1.8175, "step": 2704 }, { "epoch": 0.29233762023127635, "grad_norm": 0.34705591201782227, "learning_rate": 8e-05, "loss": 1.583, "step": 2705 }, { "epoch": 0.29244569328866316, "grad_norm": 0.3966079354286194, "learning_rate": 8e-05, "loss": 1.8207, "step": 2706 }, { "epoch": 0.2925537663460499, "grad_norm": 0.39517587423324585, "learning_rate": 8e-05, "loss": 1.6633, "step": 2707 }, { "epoch": 0.2926618394034367, "grad_norm": 0.40129369497299194, "learning_rate": 8e-05, "loss": 1.8315, "step": 2708 }, { "epoch": 0.2927699124608235, "grad_norm": 0.37463781237602234, "learning_rate": 8e-05, "loss": 1.5864, "step": 2709 }, { "epoch": 0.2928779855182103, "grad_norm": 0.3682762384414673, "learning_rate": 8e-05, "loss": 1.6139, "step": 2710 }, { "epoch": 0.2929860585755971, "grad_norm": 0.398734450340271, "learning_rate": 8e-05, "loss": 1.8334, "step": 2711 }, { "epoch": 0.2930941316329839, "grad_norm": 0.3927443325519562, "learning_rate": 8e-05, "loss": 1.8187, "step": 2712 }, { "epoch": 0.29320220469037067, "grad_norm": 0.39369529485702515, "learning_rate": 8e-05, "loss": 1.9222, "step": 2713 }, { "epoch": 0.2933102777477575, "grad_norm": 0.4231705367565155, "learning_rate": 8e-05, "loss": 1.7626, "step": 2714 }, { "epoch": 0.2934183508051443, "grad_norm": 0.34941771626472473, "learning_rate": 8e-05, "loss": 1.7377, "step": 2715 }, { "epoch": 0.2935264238625311, "grad_norm": 0.371358722448349, "learning_rate": 8e-05, "loss": 1.7137, "step": 2716 }, { "epoch": 0.2936344969199179, "grad_norm": 0.34836313128471375, "learning_rate": 8e-05, "loss": 1.6093, "step": 2717 }, { "epoch": 0.29374256997730464, "grad_norm": 0.3492712676525116, "learning_rate": 8e-05, "loss": 1.7237, "step": 2718 }, { "epoch": 0.29385064303469144, "grad_norm": 0.36492788791656494, "learning_rate": 8e-05, "loss": 1.6153, "step": 2719 }, { "epoch": 0.29395871609207824, "grad_norm": 0.3967876434326172, "learning_rate": 8e-05, "loss": 1.8939, "step": 2720 }, { "epoch": 0.29406678914946505, "grad_norm": 0.39690834283828735, "learning_rate": 8e-05, "loss": 1.6501, "step": 2721 }, { "epoch": 0.29417486220685185, "grad_norm": 0.3604986071586609, "learning_rate": 8e-05, "loss": 1.7666, "step": 2722 }, { "epoch": 0.2942829352642386, "grad_norm": 0.3591908812522888, "learning_rate": 8e-05, "loss": 1.6981, "step": 2723 }, { "epoch": 0.2943910083216254, "grad_norm": 0.3666079044342041, "learning_rate": 8e-05, "loss": 1.6843, "step": 2724 }, { "epoch": 0.2944990813790122, "grad_norm": 0.39117714762687683, "learning_rate": 8e-05, "loss": 1.7832, "step": 2725 }, { "epoch": 0.294607154436399, "grad_norm": 0.36729124188423157, "learning_rate": 8e-05, "loss": 1.7226, "step": 2726 }, { "epoch": 0.2947152274937858, "grad_norm": 0.3750307559967041, "learning_rate": 8e-05, "loss": 1.8592, "step": 2727 }, { "epoch": 0.2948233005511726, "grad_norm": 0.3781711757183075, "learning_rate": 8e-05, "loss": 1.7022, "step": 2728 }, { "epoch": 0.29493137360855937, "grad_norm": 0.3596431016921997, "learning_rate": 8e-05, "loss": 1.739, "step": 2729 }, { "epoch": 0.29503944666594617, "grad_norm": 0.37114667892456055, "learning_rate": 8e-05, "loss": 1.6202, "step": 2730 }, { "epoch": 0.295147519723333, "grad_norm": 0.3953438997268677, "learning_rate": 8e-05, "loss": 1.7305, "step": 2731 }, { "epoch": 0.2952555927807198, "grad_norm": 0.36601582169532776, "learning_rate": 8e-05, "loss": 1.6548, "step": 2732 }, { "epoch": 0.2953636658381066, "grad_norm": 0.3728499412536621, "learning_rate": 8e-05, "loss": 1.7055, "step": 2733 }, { "epoch": 0.29547173889549333, "grad_norm": 0.4207839071750641, "learning_rate": 8e-05, "loss": 1.8033, "step": 2734 }, { "epoch": 0.29557981195288013, "grad_norm": 0.3564106822013855, "learning_rate": 8e-05, "loss": 1.6848, "step": 2735 }, { "epoch": 0.29568788501026694, "grad_norm": 0.393136203289032, "learning_rate": 8e-05, "loss": 1.8076, "step": 2736 }, { "epoch": 0.29579595806765374, "grad_norm": 0.3664186894893646, "learning_rate": 8e-05, "loss": 1.6344, "step": 2737 }, { "epoch": 0.29590403112504055, "grad_norm": 0.40782058238983154, "learning_rate": 8e-05, "loss": 1.9878, "step": 2738 }, { "epoch": 0.2960121041824273, "grad_norm": 0.45472556352615356, "learning_rate": 8e-05, "loss": 2.0088, "step": 2739 }, { "epoch": 0.2961201772398141, "grad_norm": 0.40115949511528015, "learning_rate": 8e-05, "loss": 1.7765, "step": 2740 }, { "epoch": 0.2962282502972009, "grad_norm": 0.39794114232063293, "learning_rate": 8e-05, "loss": 1.724, "step": 2741 }, { "epoch": 0.2963363233545877, "grad_norm": 0.36665233969688416, "learning_rate": 8e-05, "loss": 1.7203, "step": 2742 }, { "epoch": 0.2964443964119745, "grad_norm": 0.4037715494632721, "learning_rate": 8e-05, "loss": 1.611, "step": 2743 }, { "epoch": 0.2965524694693613, "grad_norm": 0.396152526140213, "learning_rate": 8e-05, "loss": 1.8993, "step": 2744 }, { "epoch": 0.29666054252674806, "grad_norm": 0.37332990765571594, "learning_rate": 8e-05, "loss": 1.6967, "step": 2745 }, { "epoch": 0.29676861558413486, "grad_norm": 0.4242761433124542, "learning_rate": 8e-05, "loss": 1.8793, "step": 2746 }, { "epoch": 0.29687668864152167, "grad_norm": 0.35311606526374817, "learning_rate": 8e-05, "loss": 1.7552, "step": 2747 }, { "epoch": 0.2969847616989085, "grad_norm": 0.3555096983909607, "learning_rate": 8e-05, "loss": 1.5858, "step": 2748 }, { "epoch": 0.2970928347562953, "grad_norm": 0.3928786814212799, "learning_rate": 8e-05, "loss": 1.8462, "step": 2749 }, { "epoch": 0.297200907813682, "grad_norm": 0.40916478633880615, "learning_rate": 8e-05, "loss": 1.9759, "step": 2750 }, { "epoch": 0.29730898087106883, "grad_norm": 0.36609891057014465, "learning_rate": 8e-05, "loss": 1.7233, "step": 2751 }, { "epoch": 0.29741705392845563, "grad_norm": 0.37210315465927124, "learning_rate": 8e-05, "loss": 1.806, "step": 2752 }, { "epoch": 0.29752512698584244, "grad_norm": 0.3899574279785156, "learning_rate": 8e-05, "loss": 1.9027, "step": 2753 }, { "epoch": 0.29763320004322924, "grad_norm": 0.3988085389137268, "learning_rate": 8e-05, "loss": 1.6155, "step": 2754 }, { "epoch": 0.29774127310061604, "grad_norm": 0.3713085651397705, "learning_rate": 8e-05, "loss": 1.8582, "step": 2755 }, { "epoch": 0.2978493461580028, "grad_norm": 0.3908194303512573, "learning_rate": 8e-05, "loss": 1.7409, "step": 2756 }, { "epoch": 0.2979574192153896, "grad_norm": 0.41051924228668213, "learning_rate": 8e-05, "loss": 1.668, "step": 2757 }, { "epoch": 0.2980654922727764, "grad_norm": 0.3800656795501709, "learning_rate": 8e-05, "loss": 1.8168, "step": 2758 }, { "epoch": 0.2981735653301632, "grad_norm": 0.38788944482803345, "learning_rate": 8e-05, "loss": 1.7245, "step": 2759 }, { "epoch": 0.29828163838755, "grad_norm": 0.3720516860485077, "learning_rate": 8e-05, "loss": 1.8035, "step": 2760 }, { "epoch": 0.29838971144493676, "grad_norm": 0.3911111354827881, "learning_rate": 8e-05, "loss": 1.8724, "step": 2761 }, { "epoch": 0.29849778450232356, "grad_norm": 0.39373984932899475, "learning_rate": 8e-05, "loss": 1.7231, "step": 2762 }, { "epoch": 0.29860585755971036, "grad_norm": 0.3459130525588989, "learning_rate": 8e-05, "loss": 1.6581, "step": 2763 }, { "epoch": 0.29871393061709717, "grad_norm": 0.39422255754470825, "learning_rate": 8e-05, "loss": 1.6024, "step": 2764 }, { "epoch": 0.29882200367448397, "grad_norm": 0.3514842391014099, "learning_rate": 8e-05, "loss": 1.6137, "step": 2765 }, { "epoch": 0.2989300767318707, "grad_norm": 0.3713415563106537, "learning_rate": 8e-05, "loss": 1.7875, "step": 2766 }, { "epoch": 0.2990381497892575, "grad_norm": 0.38729971647262573, "learning_rate": 8e-05, "loss": 1.6968, "step": 2767 }, { "epoch": 0.2991462228466443, "grad_norm": 0.36966803669929504, "learning_rate": 8e-05, "loss": 1.6488, "step": 2768 }, { "epoch": 0.29925429590403113, "grad_norm": 0.3689405620098114, "learning_rate": 8e-05, "loss": 1.781, "step": 2769 }, { "epoch": 0.29936236896141794, "grad_norm": 0.3855065405368805, "learning_rate": 8e-05, "loss": 1.7771, "step": 2770 }, { "epoch": 0.29947044201880474, "grad_norm": 0.36936676502227783, "learning_rate": 8e-05, "loss": 1.7675, "step": 2771 }, { "epoch": 0.2995785150761915, "grad_norm": 0.39187923073768616, "learning_rate": 8e-05, "loss": 1.7893, "step": 2772 }, { "epoch": 0.2996865881335783, "grad_norm": 0.369331955909729, "learning_rate": 8e-05, "loss": 1.7343, "step": 2773 }, { "epoch": 0.2997946611909651, "grad_norm": 0.38518908619880676, "learning_rate": 8e-05, "loss": 1.6393, "step": 2774 }, { "epoch": 0.2999027342483519, "grad_norm": 0.382506787776947, "learning_rate": 8e-05, "loss": 1.7603, "step": 2775 }, { "epoch": 0.3000108073057387, "grad_norm": 0.5179709792137146, "learning_rate": 8e-05, "loss": 2.0043, "step": 2776 }, { "epoch": 0.30011888036312545, "grad_norm": 0.4107379913330078, "learning_rate": 8e-05, "loss": 1.6966, "step": 2777 }, { "epoch": 0.30022695342051225, "grad_norm": 0.41739919781684875, "learning_rate": 8e-05, "loss": 1.8737, "step": 2778 }, { "epoch": 0.30033502647789906, "grad_norm": 0.4223654270172119, "learning_rate": 8e-05, "loss": 1.8622, "step": 2779 }, { "epoch": 0.30044309953528586, "grad_norm": 0.38513776659965515, "learning_rate": 8e-05, "loss": 1.7186, "step": 2780 }, { "epoch": 0.30055117259267267, "grad_norm": 0.35849669575691223, "learning_rate": 8e-05, "loss": 1.7297, "step": 2781 }, { "epoch": 0.3006592456500594, "grad_norm": 0.42787253856658936, "learning_rate": 8e-05, "loss": 1.9259, "step": 2782 }, { "epoch": 0.3007673187074462, "grad_norm": 0.3671242296695709, "learning_rate": 8e-05, "loss": 1.6095, "step": 2783 }, { "epoch": 0.300875391764833, "grad_norm": 0.38985350728034973, "learning_rate": 8e-05, "loss": 1.8322, "step": 2784 }, { "epoch": 0.3009834648222198, "grad_norm": 0.37536555528640747, "learning_rate": 8e-05, "loss": 1.8342, "step": 2785 }, { "epoch": 0.30109153787960663, "grad_norm": 0.3566596508026123, "learning_rate": 8e-05, "loss": 1.6076, "step": 2786 }, { "epoch": 0.30119961093699343, "grad_norm": 0.38576146960258484, "learning_rate": 8e-05, "loss": 1.5171, "step": 2787 }, { "epoch": 0.3013076839943802, "grad_norm": 0.3565886616706848, "learning_rate": 8e-05, "loss": 1.6998, "step": 2788 }, { "epoch": 0.301415757051767, "grad_norm": 0.3559359312057495, "learning_rate": 8e-05, "loss": 1.7064, "step": 2789 }, { "epoch": 0.3015238301091538, "grad_norm": 0.35267412662506104, "learning_rate": 8e-05, "loss": 1.645, "step": 2790 }, { "epoch": 0.3016319031665406, "grad_norm": 0.39178356528282166, "learning_rate": 8e-05, "loss": 1.8374, "step": 2791 }, { "epoch": 0.3017399762239274, "grad_norm": 0.37979385256767273, "learning_rate": 8e-05, "loss": 1.8402, "step": 2792 }, { "epoch": 0.30184804928131415, "grad_norm": 0.3830778896808624, "learning_rate": 8e-05, "loss": 1.7141, "step": 2793 }, { "epoch": 0.30195612233870095, "grad_norm": 0.3955959379673004, "learning_rate": 8e-05, "loss": 1.8396, "step": 2794 }, { "epoch": 0.30206419539608775, "grad_norm": 0.3882089853286743, "learning_rate": 8e-05, "loss": 1.7059, "step": 2795 }, { "epoch": 0.30217226845347456, "grad_norm": 0.3677847981452942, "learning_rate": 8e-05, "loss": 1.7654, "step": 2796 }, { "epoch": 0.30228034151086136, "grad_norm": 0.3848576545715332, "learning_rate": 8e-05, "loss": 1.7194, "step": 2797 }, { "epoch": 0.3023884145682481, "grad_norm": 0.3704456388950348, "learning_rate": 8e-05, "loss": 1.793, "step": 2798 }, { "epoch": 0.3024964876256349, "grad_norm": 0.3719424307346344, "learning_rate": 8e-05, "loss": 1.7165, "step": 2799 }, { "epoch": 0.3026045606830217, "grad_norm": 0.39718568325042725, "learning_rate": 8e-05, "loss": 1.8071, "step": 2800 }, { "epoch": 0.3027126337404085, "grad_norm": 0.3685368597507477, "learning_rate": 8e-05, "loss": 1.785, "step": 2801 }, { "epoch": 0.3028207067977953, "grad_norm": 0.4045741558074951, "learning_rate": 8e-05, "loss": 1.7183, "step": 2802 }, { "epoch": 0.30292877985518213, "grad_norm": 0.38734057545661926, "learning_rate": 8e-05, "loss": 1.7426, "step": 2803 }, { "epoch": 0.3030368529125689, "grad_norm": 0.40937939286231995, "learning_rate": 8e-05, "loss": 1.9377, "step": 2804 }, { "epoch": 0.3031449259699557, "grad_norm": 0.41106492280960083, "learning_rate": 8e-05, "loss": 1.7681, "step": 2805 }, { "epoch": 0.3032529990273425, "grad_norm": 0.3860428035259247, "learning_rate": 8e-05, "loss": 1.7771, "step": 2806 }, { "epoch": 0.3033610720847293, "grad_norm": 0.38663002848625183, "learning_rate": 8e-05, "loss": 1.6858, "step": 2807 }, { "epoch": 0.3034691451421161, "grad_norm": 0.40180304646492004, "learning_rate": 8e-05, "loss": 1.7658, "step": 2808 }, { "epoch": 0.30357721819950284, "grad_norm": 0.437874436378479, "learning_rate": 8e-05, "loss": 1.7222, "step": 2809 }, { "epoch": 0.30368529125688964, "grad_norm": 0.39362409710884094, "learning_rate": 8e-05, "loss": 1.8068, "step": 2810 }, { "epoch": 0.30379336431427645, "grad_norm": 0.41877835988998413, "learning_rate": 8e-05, "loss": 1.9216, "step": 2811 }, { "epoch": 0.30390143737166325, "grad_norm": 0.3777587115764618, "learning_rate": 8e-05, "loss": 1.7634, "step": 2812 }, { "epoch": 0.30400951042905006, "grad_norm": 0.38361722230911255, "learning_rate": 8e-05, "loss": 1.7395, "step": 2813 }, { "epoch": 0.3041175834864368, "grad_norm": 0.3810099959373474, "learning_rate": 8e-05, "loss": 1.8252, "step": 2814 }, { "epoch": 0.3042256565438236, "grad_norm": 0.385941743850708, "learning_rate": 8e-05, "loss": 1.8048, "step": 2815 }, { "epoch": 0.3043337296012104, "grad_norm": 0.3661251664161682, "learning_rate": 8e-05, "loss": 1.5477, "step": 2816 }, { "epoch": 0.3044418026585972, "grad_norm": 0.4016244113445282, "learning_rate": 8e-05, "loss": 1.7684, "step": 2817 }, { "epoch": 0.304549875715984, "grad_norm": 0.37685319781303406, "learning_rate": 8e-05, "loss": 1.6606, "step": 2818 }, { "epoch": 0.3046579487733708, "grad_norm": 0.3665071725845337, "learning_rate": 8e-05, "loss": 1.8231, "step": 2819 }, { "epoch": 0.30476602183075757, "grad_norm": 0.3693774938583374, "learning_rate": 8e-05, "loss": 1.7373, "step": 2820 }, { "epoch": 0.3048740948881444, "grad_norm": 0.3791259229183197, "learning_rate": 8e-05, "loss": 1.8299, "step": 2821 }, { "epoch": 0.3049821679455312, "grad_norm": 0.35839998722076416, "learning_rate": 8e-05, "loss": 1.6922, "step": 2822 }, { "epoch": 0.305090241002918, "grad_norm": 0.38520944118499756, "learning_rate": 8e-05, "loss": 1.7927, "step": 2823 }, { "epoch": 0.3051983140603048, "grad_norm": 0.3816578686237335, "learning_rate": 8e-05, "loss": 1.7264, "step": 2824 }, { "epoch": 0.30530638711769154, "grad_norm": 0.38569384813308716, "learning_rate": 8e-05, "loss": 1.8728, "step": 2825 }, { "epoch": 0.30541446017507834, "grad_norm": 0.370116651058197, "learning_rate": 8e-05, "loss": 1.7284, "step": 2826 }, { "epoch": 0.30552253323246514, "grad_norm": 0.36661067605018616, "learning_rate": 8e-05, "loss": 1.6363, "step": 2827 }, { "epoch": 0.30563060628985195, "grad_norm": 0.3710159957408905, "learning_rate": 8e-05, "loss": 1.8074, "step": 2828 }, { "epoch": 0.30573867934723875, "grad_norm": 0.3679962158203125, "learning_rate": 8e-05, "loss": 1.8645, "step": 2829 }, { "epoch": 0.3058467524046255, "grad_norm": 0.41197913885116577, "learning_rate": 8e-05, "loss": 1.8282, "step": 2830 }, { "epoch": 0.3059548254620123, "grad_norm": 0.37579745054244995, "learning_rate": 8e-05, "loss": 1.6248, "step": 2831 }, { "epoch": 0.3060628985193991, "grad_norm": 0.3733593225479126, "learning_rate": 8e-05, "loss": 1.6092, "step": 2832 }, { "epoch": 0.3061709715767859, "grad_norm": 0.36421963572502136, "learning_rate": 8e-05, "loss": 1.7618, "step": 2833 }, { "epoch": 0.3062790446341727, "grad_norm": 0.3788909614086151, "learning_rate": 8e-05, "loss": 1.8222, "step": 2834 }, { "epoch": 0.3063871176915595, "grad_norm": 0.4105842709541321, "learning_rate": 8e-05, "loss": 1.7551, "step": 2835 }, { "epoch": 0.30649519074894627, "grad_norm": 0.3690252900123596, "learning_rate": 8e-05, "loss": 1.7652, "step": 2836 }, { "epoch": 0.30660326380633307, "grad_norm": 0.38524535298347473, "learning_rate": 8e-05, "loss": 1.7086, "step": 2837 }, { "epoch": 0.3067113368637199, "grad_norm": 0.35578757524490356, "learning_rate": 8e-05, "loss": 1.6818, "step": 2838 }, { "epoch": 0.3068194099211067, "grad_norm": 0.4109478294849396, "learning_rate": 8e-05, "loss": 1.6761, "step": 2839 }, { "epoch": 0.3069274829784935, "grad_norm": 0.3837120532989502, "learning_rate": 8e-05, "loss": 1.5612, "step": 2840 }, { "epoch": 0.30703555603588023, "grad_norm": 0.3675451874732971, "learning_rate": 8e-05, "loss": 1.818, "step": 2841 }, { "epoch": 0.30714362909326703, "grad_norm": 0.3980099558830261, "learning_rate": 8e-05, "loss": 1.9134, "step": 2842 }, { "epoch": 0.30725170215065384, "grad_norm": 0.36603352427482605, "learning_rate": 8e-05, "loss": 1.7407, "step": 2843 }, { "epoch": 0.30735977520804064, "grad_norm": 0.3835507035255432, "learning_rate": 8e-05, "loss": 1.7966, "step": 2844 }, { "epoch": 0.30746784826542745, "grad_norm": 0.3695836663246155, "learning_rate": 8e-05, "loss": 1.5319, "step": 2845 }, { "epoch": 0.30757592132281425, "grad_norm": 0.3849472403526306, "learning_rate": 8e-05, "loss": 1.8283, "step": 2846 }, { "epoch": 0.307683994380201, "grad_norm": 0.4207453727722168, "learning_rate": 8e-05, "loss": 1.8893, "step": 2847 }, { "epoch": 0.3077920674375878, "grad_norm": 0.395069420337677, "learning_rate": 8e-05, "loss": 1.766, "step": 2848 }, { "epoch": 0.3079001404949746, "grad_norm": 0.4042578637599945, "learning_rate": 8e-05, "loss": 1.8761, "step": 2849 }, { "epoch": 0.3080082135523614, "grad_norm": 0.37202179431915283, "learning_rate": 8e-05, "loss": 1.8068, "step": 2850 }, { "epoch": 0.3081162866097482, "grad_norm": 0.3904436528682709, "learning_rate": 8e-05, "loss": 1.8431, "step": 2851 }, { "epoch": 0.30822435966713496, "grad_norm": 0.44059136509895325, "learning_rate": 8e-05, "loss": 1.8889, "step": 2852 }, { "epoch": 0.30833243272452177, "grad_norm": 0.39557191729545593, "learning_rate": 8e-05, "loss": 1.5656, "step": 2853 }, { "epoch": 0.30844050578190857, "grad_norm": 0.39514783024787903, "learning_rate": 8e-05, "loss": 1.7284, "step": 2854 }, { "epoch": 0.3085485788392954, "grad_norm": 0.3819649815559387, "learning_rate": 8e-05, "loss": 1.7746, "step": 2855 }, { "epoch": 0.3086566518966822, "grad_norm": 0.3580262064933777, "learning_rate": 8e-05, "loss": 1.7425, "step": 2856 }, { "epoch": 0.3087647249540689, "grad_norm": 0.367521733045578, "learning_rate": 8e-05, "loss": 1.8224, "step": 2857 }, { "epoch": 0.30887279801145573, "grad_norm": 0.35109928250312805, "learning_rate": 8e-05, "loss": 1.5895, "step": 2858 }, { "epoch": 0.30898087106884253, "grad_norm": 0.4035426676273346, "learning_rate": 8e-05, "loss": 1.931, "step": 2859 }, { "epoch": 0.30908894412622934, "grad_norm": 0.41659969091415405, "learning_rate": 8e-05, "loss": 1.906, "step": 2860 }, { "epoch": 0.30919701718361614, "grad_norm": 0.4266480803489685, "learning_rate": 8e-05, "loss": 1.8134, "step": 2861 }, { "epoch": 0.30930509024100294, "grad_norm": 0.38545137643814087, "learning_rate": 8e-05, "loss": 1.7012, "step": 2862 }, { "epoch": 0.3094131632983897, "grad_norm": 0.41232702136039734, "learning_rate": 8e-05, "loss": 1.9435, "step": 2863 }, { "epoch": 0.3095212363557765, "grad_norm": 0.365920752286911, "learning_rate": 8e-05, "loss": 1.7095, "step": 2864 }, { "epoch": 0.3096293094131633, "grad_norm": 0.38208508491516113, "learning_rate": 8e-05, "loss": 1.7408, "step": 2865 }, { "epoch": 0.3097373824705501, "grad_norm": 0.3837342858314514, "learning_rate": 8e-05, "loss": 1.5845, "step": 2866 }, { "epoch": 0.3098454555279369, "grad_norm": 0.3689322769641876, "learning_rate": 8e-05, "loss": 1.7638, "step": 2867 }, { "epoch": 0.30995352858532366, "grad_norm": 0.36687833070755005, "learning_rate": 8e-05, "loss": 1.7149, "step": 2868 }, { "epoch": 0.31006160164271046, "grad_norm": 0.4071006178855896, "learning_rate": 8e-05, "loss": 1.9323, "step": 2869 }, { "epoch": 0.31016967470009726, "grad_norm": 0.36862632632255554, "learning_rate": 8e-05, "loss": 1.7744, "step": 2870 }, { "epoch": 0.31027774775748407, "grad_norm": 0.3692541718482971, "learning_rate": 8e-05, "loss": 1.8932, "step": 2871 }, { "epoch": 0.31038582081487087, "grad_norm": 0.41158992052078247, "learning_rate": 8e-05, "loss": 1.8115, "step": 2872 }, { "epoch": 0.3104938938722576, "grad_norm": 0.36422809958457947, "learning_rate": 8e-05, "loss": 1.7177, "step": 2873 }, { "epoch": 0.3106019669296444, "grad_norm": 0.4066080152988434, "learning_rate": 8e-05, "loss": 1.9813, "step": 2874 }, { "epoch": 0.31071003998703123, "grad_norm": 0.43168526887893677, "learning_rate": 8e-05, "loss": 1.9608, "step": 2875 }, { "epoch": 0.31081811304441803, "grad_norm": 0.39238473773002625, "learning_rate": 8e-05, "loss": 1.8645, "step": 2876 }, { "epoch": 0.31092618610180484, "grad_norm": 0.38773417472839355, "learning_rate": 8e-05, "loss": 1.8767, "step": 2877 }, { "epoch": 0.31103425915919164, "grad_norm": 0.38922005891799927, "learning_rate": 8e-05, "loss": 1.739, "step": 2878 }, { "epoch": 0.3111423322165784, "grad_norm": 0.35113298892974854, "learning_rate": 8e-05, "loss": 1.6152, "step": 2879 }, { "epoch": 0.3112504052739652, "grad_norm": 0.39437413215637207, "learning_rate": 8e-05, "loss": 1.9242, "step": 2880 }, { "epoch": 0.311358478331352, "grad_norm": 0.39073893427848816, "learning_rate": 8e-05, "loss": 1.9062, "step": 2881 }, { "epoch": 0.3114665513887388, "grad_norm": 0.3760479688644409, "learning_rate": 8e-05, "loss": 1.7592, "step": 2882 }, { "epoch": 0.3115746244461256, "grad_norm": 0.38166531920433044, "learning_rate": 8e-05, "loss": 1.716, "step": 2883 }, { "epoch": 0.31168269750351235, "grad_norm": 0.3855013847351074, "learning_rate": 8e-05, "loss": 1.479, "step": 2884 }, { "epoch": 0.31179077056089916, "grad_norm": 0.41183751821517944, "learning_rate": 8e-05, "loss": 1.8228, "step": 2885 }, { "epoch": 0.31189884361828596, "grad_norm": 0.399077832698822, "learning_rate": 8e-05, "loss": 1.828, "step": 2886 }, { "epoch": 0.31200691667567276, "grad_norm": 0.3941633403301239, "learning_rate": 8e-05, "loss": 1.77, "step": 2887 }, { "epoch": 0.31211498973305957, "grad_norm": 0.374554842710495, "learning_rate": 8e-05, "loss": 1.6653, "step": 2888 }, { "epoch": 0.3122230627904463, "grad_norm": 0.3725683093070984, "learning_rate": 8e-05, "loss": 1.648, "step": 2889 }, { "epoch": 0.3123311358478331, "grad_norm": 0.3631357252597809, "learning_rate": 8e-05, "loss": 1.5924, "step": 2890 }, { "epoch": 0.3124392089052199, "grad_norm": 0.39798495173454285, "learning_rate": 8e-05, "loss": 1.7096, "step": 2891 }, { "epoch": 0.3125472819626067, "grad_norm": 0.39654138684272766, "learning_rate": 8e-05, "loss": 1.7855, "step": 2892 }, { "epoch": 0.31265535501999353, "grad_norm": 0.4354475736618042, "learning_rate": 8e-05, "loss": 1.9467, "step": 2893 }, { "epoch": 0.31276342807738033, "grad_norm": 0.4001348316669464, "learning_rate": 8e-05, "loss": 1.866, "step": 2894 }, { "epoch": 0.3128715011347671, "grad_norm": 0.3934697210788727, "learning_rate": 8e-05, "loss": 1.7055, "step": 2895 }, { "epoch": 0.3129795741921539, "grad_norm": 0.4336510896682739, "learning_rate": 8e-05, "loss": 1.8726, "step": 2896 }, { "epoch": 0.3130876472495407, "grad_norm": 0.35438022017478943, "learning_rate": 8e-05, "loss": 1.7279, "step": 2897 }, { "epoch": 0.3131957203069275, "grad_norm": 0.40890607237815857, "learning_rate": 8e-05, "loss": 1.7566, "step": 2898 }, { "epoch": 0.3133037933643143, "grad_norm": 0.41066208481788635, "learning_rate": 8e-05, "loss": 1.7447, "step": 2899 }, { "epoch": 0.31341186642170105, "grad_norm": 0.368725448846817, "learning_rate": 8e-05, "loss": 1.6884, "step": 2900 }, { "epoch": 0.31351993947908785, "grad_norm": 0.4108510911464691, "learning_rate": 8e-05, "loss": 1.9705, "step": 2901 }, { "epoch": 0.31362801253647465, "grad_norm": 0.3954737186431885, "learning_rate": 8e-05, "loss": 1.7379, "step": 2902 }, { "epoch": 0.31373608559386146, "grad_norm": 0.44700920581817627, "learning_rate": 8e-05, "loss": 1.7883, "step": 2903 }, { "epoch": 0.31384415865124826, "grad_norm": 0.33998164534568787, "learning_rate": 8e-05, "loss": 1.5107, "step": 2904 }, { "epoch": 0.313952231708635, "grad_norm": 0.35271117091178894, "learning_rate": 8e-05, "loss": 1.6449, "step": 2905 }, { "epoch": 0.3140603047660218, "grad_norm": 0.4002251923084259, "learning_rate": 8e-05, "loss": 1.7967, "step": 2906 }, { "epoch": 0.3141683778234086, "grad_norm": 0.40751349925994873, "learning_rate": 8e-05, "loss": 1.726, "step": 2907 }, { "epoch": 0.3142764508807954, "grad_norm": 0.38629281520843506, "learning_rate": 8e-05, "loss": 1.7872, "step": 2908 }, { "epoch": 0.3143845239381822, "grad_norm": 0.38523009419441223, "learning_rate": 8e-05, "loss": 1.8642, "step": 2909 }, { "epoch": 0.31449259699556903, "grad_norm": 0.4356428384780884, "learning_rate": 8e-05, "loss": 1.8041, "step": 2910 }, { "epoch": 0.3146006700529558, "grad_norm": 0.386976957321167, "learning_rate": 8e-05, "loss": 1.731, "step": 2911 }, { "epoch": 0.3147087431103426, "grad_norm": 0.4100055992603302, "learning_rate": 8e-05, "loss": 1.8026, "step": 2912 }, { "epoch": 0.3148168161677294, "grad_norm": 0.3868879973888397, "learning_rate": 8e-05, "loss": 1.5354, "step": 2913 }, { "epoch": 0.3149248892251162, "grad_norm": 0.359494686126709, "learning_rate": 8e-05, "loss": 1.6117, "step": 2914 }, { "epoch": 0.315032962282503, "grad_norm": 0.3833145201206207, "learning_rate": 8e-05, "loss": 1.7569, "step": 2915 }, { "epoch": 0.31514103533988974, "grad_norm": 0.4158884584903717, "learning_rate": 8e-05, "loss": 1.7693, "step": 2916 }, { "epoch": 0.31524910839727655, "grad_norm": 0.41443008184432983, "learning_rate": 8e-05, "loss": 1.6618, "step": 2917 }, { "epoch": 0.31535718145466335, "grad_norm": 0.35206183791160583, "learning_rate": 8e-05, "loss": 1.5787, "step": 2918 }, { "epoch": 0.31546525451205015, "grad_norm": 0.3966812491416931, "learning_rate": 8e-05, "loss": 1.6348, "step": 2919 }, { "epoch": 0.31557332756943696, "grad_norm": 0.39897847175598145, "learning_rate": 8e-05, "loss": 1.658, "step": 2920 }, { "epoch": 0.31568140062682376, "grad_norm": 0.3685523271560669, "learning_rate": 8e-05, "loss": 1.6447, "step": 2921 }, { "epoch": 0.3157894736842105, "grad_norm": 0.37183451652526855, "learning_rate": 8e-05, "loss": 1.5332, "step": 2922 }, { "epoch": 0.3158975467415973, "grad_norm": 0.3534536063671112, "learning_rate": 8e-05, "loss": 1.5273, "step": 2923 }, { "epoch": 0.3160056197989841, "grad_norm": 0.36801114678382874, "learning_rate": 8e-05, "loss": 1.8019, "step": 2924 }, { "epoch": 0.3161136928563709, "grad_norm": 0.3689340353012085, "learning_rate": 8e-05, "loss": 1.6638, "step": 2925 }, { "epoch": 0.3162217659137577, "grad_norm": 0.37881961464881897, "learning_rate": 8e-05, "loss": 1.7105, "step": 2926 }, { "epoch": 0.3163298389711445, "grad_norm": 0.37032562494277954, "learning_rate": 8e-05, "loss": 1.7243, "step": 2927 }, { "epoch": 0.3164379120285313, "grad_norm": 0.40114015340805054, "learning_rate": 8e-05, "loss": 1.814, "step": 2928 }, { "epoch": 0.3165459850859181, "grad_norm": 0.4087194502353668, "learning_rate": 8e-05, "loss": 1.7618, "step": 2929 }, { "epoch": 0.3166540581433049, "grad_norm": 0.44881680607795715, "learning_rate": 8e-05, "loss": 1.8863, "step": 2930 }, { "epoch": 0.3167621312006917, "grad_norm": 0.3572789430618286, "learning_rate": 8e-05, "loss": 1.7184, "step": 2931 }, { "epoch": 0.31687020425807844, "grad_norm": 0.3730814754962921, "learning_rate": 8e-05, "loss": 1.7025, "step": 2932 }, { "epoch": 0.31697827731546524, "grad_norm": 0.40190377831459045, "learning_rate": 8e-05, "loss": 1.8714, "step": 2933 }, { "epoch": 0.31708635037285204, "grad_norm": 0.3877315819263458, "learning_rate": 8e-05, "loss": 1.7679, "step": 2934 }, { "epoch": 0.31719442343023885, "grad_norm": 0.35322630405426025, "learning_rate": 8e-05, "loss": 1.4556, "step": 2935 }, { "epoch": 0.31730249648762565, "grad_norm": 0.42431527376174927, "learning_rate": 8e-05, "loss": 1.8214, "step": 2936 }, { "epoch": 0.31741056954501246, "grad_norm": 0.3871442675590515, "learning_rate": 8e-05, "loss": 1.7399, "step": 2937 }, { "epoch": 0.3175186426023992, "grad_norm": 0.39737236499786377, "learning_rate": 8e-05, "loss": 1.8192, "step": 2938 }, { "epoch": 0.317626715659786, "grad_norm": 0.4183724522590637, "learning_rate": 8e-05, "loss": 1.8086, "step": 2939 }, { "epoch": 0.3177347887171728, "grad_norm": 0.38821372389793396, "learning_rate": 8e-05, "loss": 1.6937, "step": 2940 }, { "epoch": 0.3178428617745596, "grad_norm": 0.432868629693985, "learning_rate": 8e-05, "loss": 1.8571, "step": 2941 }, { "epoch": 0.3179509348319464, "grad_norm": 0.36649852991104126, "learning_rate": 8e-05, "loss": 1.5875, "step": 2942 }, { "epoch": 0.31805900788933317, "grad_norm": 0.4657868444919586, "learning_rate": 8e-05, "loss": 1.9013, "step": 2943 }, { "epoch": 0.31816708094671997, "grad_norm": 0.40176424384117126, "learning_rate": 8e-05, "loss": 1.901, "step": 2944 }, { "epoch": 0.3182751540041068, "grad_norm": 0.373000830411911, "learning_rate": 8e-05, "loss": 1.807, "step": 2945 }, { "epoch": 0.3183832270614936, "grad_norm": 0.37759608030319214, "learning_rate": 8e-05, "loss": 1.8529, "step": 2946 }, { "epoch": 0.3184913001188804, "grad_norm": 0.3805919885635376, "learning_rate": 8e-05, "loss": 1.6764, "step": 2947 }, { "epoch": 0.31859937317626713, "grad_norm": 0.3967913091182709, "learning_rate": 8e-05, "loss": 1.7912, "step": 2948 }, { "epoch": 0.31870744623365393, "grad_norm": 0.39837178587913513, "learning_rate": 8e-05, "loss": 1.9178, "step": 2949 }, { "epoch": 0.31881551929104074, "grad_norm": 0.35366323590278625, "learning_rate": 8e-05, "loss": 1.6997, "step": 2950 }, { "epoch": 0.31892359234842754, "grad_norm": 0.3793390393257141, "learning_rate": 8e-05, "loss": 1.4798, "step": 2951 }, { "epoch": 0.31903166540581435, "grad_norm": 0.3694332242012024, "learning_rate": 8e-05, "loss": 1.7257, "step": 2952 }, { "epoch": 0.31913973846320115, "grad_norm": 0.4321056604385376, "learning_rate": 8e-05, "loss": 1.8867, "step": 2953 }, { "epoch": 0.3192478115205879, "grad_norm": 0.37852564454078674, "learning_rate": 8e-05, "loss": 1.7001, "step": 2954 }, { "epoch": 0.3193558845779747, "grad_norm": 0.3669835925102234, "learning_rate": 8e-05, "loss": 1.6707, "step": 2955 }, { "epoch": 0.3194639576353615, "grad_norm": 0.3501361012458801, "learning_rate": 8e-05, "loss": 1.7067, "step": 2956 }, { "epoch": 0.3195720306927483, "grad_norm": 0.37453052401542664, "learning_rate": 8e-05, "loss": 1.5739, "step": 2957 }, { "epoch": 0.3196801037501351, "grad_norm": 0.40657395124435425, "learning_rate": 8e-05, "loss": 1.8, "step": 2958 }, { "epoch": 0.31978817680752186, "grad_norm": 0.40382397174835205, "learning_rate": 8e-05, "loss": 1.8852, "step": 2959 }, { "epoch": 0.31989624986490867, "grad_norm": 0.3884938657283783, "learning_rate": 8e-05, "loss": 1.6144, "step": 2960 }, { "epoch": 0.32000432292229547, "grad_norm": 0.3831300437450409, "learning_rate": 8e-05, "loss": 1.7825, "step": 2961 }, { "epoch": 0.3201123959796823, "grad_norm": 0.4609825015068054, "learning_rate": 8e-05, "loss": 1.911, "step": 2962 }, { "epoch": 0.3202204690370691, "grad_norm": 0.3565077781677246, "learning_rate": 8e-05, "loss": 1.8518, "step": 2963 }, { "epoch": 0.3203285420944558, "grad_norm": 0.35824379324913025, "learning_rate": 8e-05, "loss": 1.5757, "step": 2964 }, { "epoch": 0.32043661515184263, "grad_norm": 0.38923028111457825, "learning_rate": 8e-05, "loss": 1.8399, "step": 2965 }, { "epoch": 0.32054468820922943, "grad_norm": 0.4200632870197296, "learning_rate": 8e-05, "loss": 1.8988, "step": 2966 }, { "epoch": 0.32065276126661624, "grad_norm": 0.42543795704841614, "learning_rate": 8e-05, "loss": 1.8122, "step": 2967 }, { "epoch": 0.32076083432400304, "grad_norm": 0.3648354411125183, "learning_rate": 8e-05, "loss": 1.7121, "step": 2968 }, { "epoch": 0.32086890738138985, "grad_norm": 0.39955025911331177, "learning_rate": 8e-05, "loss": 1.9118, "step": 2969 }, { "epoch": 0.3209769804387766, "grad_norm": 0.37788957357406616, "learning_rate": 8e-05, "loss": 1.8012, "step": 2970 }, { "epoch": 0.3210850534961634, "grad_norm": 0.41708123683929443, "learning_rate": 8e-05, "loss": 1.8201, "step": 2971 }, { "epoch": 0.3211931265535502, "grad_norm": 0.39043566584587097, "learning_rate": 8e-05, "loss": 1.7285, "step": 2972 }, { "epoch": 0.321301199610937, "grad_norm": 0.3980652391910553, "learning_rate": 8e-05, "loss": 1.8051, "step": 2973 }, { "epoch": 0.3214092726683238, "grad_norm": 0.3829914629459381, "learning_rate": 8e-05, "loss": 1.7186, "step": 2974 }, { "epoch": 0.32151734572571056, "grad_norm": 0.3798570930957794, "learning_rate": 8e-05, "loss": 1.7397, "step": 2975 }, { "epoch": 0.32162541878309736, "grad_norm": 0.36752450466156006, "learning_rate": 8e-05, "loss": 1.6689, "step": 2976 }, { "epoch": 0.32173349184048416, "grad_norm": 0.3925345540046692, "learning_rate": 8e-05, "loss": 1.9337, "step": 2977 }, { "epoch": 0.32184156489787097, "grad_norm": 0.379050612449646, "learning_rate": 8e-05, "loss": 1.7869, "step": 2978 }, { "epoch": 0.3219496379552578, "grad_norm": 0.41210389137268066, "learning_rate": 8e-05, "loss": 1.9441, "step": 2979 }, { "epoch": 0.3220577110126445, "grad_norm": 0.37567928433418274, "learning_rate": 8e-05, "loss": 1.7532, "step": 2980 }, { "epoch": 0.3221657840700313, "grad_norm": 0.38562798500061035, "learning_rate": 8e-05, "loss": 1.7766, "step": 2981 }, { "epoch": 0.32227385712741813, "grad_norm": 0.36712729930877686, "learning_rate": 8e-05, "loss": 1.6827, "step": 2982 }, { "epoch": 0.32238193018480493, "grad_norm": 0.3820264935493469, "learning_rate": 8e-05, "loss": 1.8901, "step": 2983 }, { "epoch": 0.32249000324219174, "grad_norm": 0.4436863660812378, "learning_rate": 8e-05, "loss": 1.9087, "step": 2984 }, { "epoch": 0.32259807629957854, "grad_norm": 0.36913198232650757, "learning_rate": 8e-05, "loss": 1.6119, "step": 2985 }, { "epoch": 0.3227061493569653, "grad_norm": 0.40391990542411804, "learning_rate": 8e-05, "loss": 1.7386, "step": 2986 }, { "epoch": 0.3228142224143521, "grad_norm": 0.40781503915786743, "learning_rate": 8e-05, "loss": 1.7348, "step": 2987 }, { "epoch": 0.3229222954717389, "grad_norm": 0.39497804641723633, "learning_rate": 8e-05, "loss": 1.7755, "step": 2988 }, { "epoch": 0.3230303685291257, "grad_norm": 0.3795987069606781, "learning_rate": 8e-05, "loss": 1.7753, "step": 2989 }, { "epoch": 0.3231384415865125, "grad_norm": 0.38672304153442383, "learning_rate": 8e-05, "loss": 1.7625, "step": 2990 }, { "epoch": 0.32324651464389925, "grad_norm": 0.36739087104797363, "learning_rate": 8e-05, "loss": 1.7472, "step": 2991 }, { "epoch": 0.32335458770128606, "grad_norm": 0.3887999355792999, "learning_rate": 8e-05, "loss": 1.662, "step": 2992 }, { "epoch": 0.32346266075867286, "grad_norm": 0.36788210272789, "learning_rate": 8e-05, "loss": 1.7344, "step": 2993 }, { "epoch": 0.32357073381605966, "grad_norm": 0.37225428223609924, "learning_rate": 8e-05, "loss": 1.7525, "step": 2994 }, { "epoch": 0.32367880687344647, "grad_norm": 0.39085617661476135, "learning_rate": 8e-05, "loss": 1.7446, "step": 2995 }, { "epoch": 0.3237868799308332, "grad_norm": 0.3666759431362152, "learning_rate": 8e-05, "loss": 1.641, "step": 2996 }, { "epoch": 0.32389495298822, "grad_norm": 0.35601726174354553, "learning_rate": 8e-05, "loss": 1.7032, "step": 2997 }, { "epoch": 0.3240030260456068, "grad_norm": 0.3973410427570343, "learning_rate": 8e-05, "loss": 1.765, "step": 2998 }, { "epoch": 0.3241110991029936, "grad_norm": 0.3608280420303345, "learning_rate": 8e-05, "loss": 1.7558, "step": 2999 }, { "epoch": 0.32421917216038043, "grad_norm": 0.3686004877090454, "learning_rate": 8e-05, "loss": 1.6906, "step": 3000 }, { "epoch": 0.32432724521776723, "grad_norm": 0.3809454143047333, "learning_rate": 8e-05, "loss": 1.7286, "step": 3001 }, { "epoch": 0.324435318275154, "grad_norm": 0.4698570966720581, "learning_rate": 8e-05, "loss": 1.9005, "step": 3002 }, { "epoch": 0.3245433913325408, "grad_norm": 0.38386544585227966, "learning_rate": 8e-05, "loss": 1.5026, "step": 3003 }, { "epoch": 0.3246514643899276, "grad_norm": 0.400448739528656, "learning_rate": 8e-05, "loss": 1.8962, "step": 3004 }, { "epoch": 0.3247595374473144, "grad_norm": 0.38756120204925537, "learning_rate": 8e-05, "loss": 1.6806, "step": 3005 }, { "epoch": 0.3248676105047012, "grad_norm": 0.38079118728637695, "learning_rate": 8e-05, "loss": 1.8685, "step": 3006 }, { "epoch": 0.32497568356208795, "grad_norm": 0.38144582509994507, "learning_rate": 8e-05, "loss": 1.4927, "step": 3007 }, { "epoch": 0.32508375661947475, "grad_norm": 0.46014514565467834, "learning_rate": 8e-05, "loss": 1.8318, "step": 3008 }, { "epoch": 0.32519182967686155, "grad_norm": 0.3712429702281952, "learning_rate": 8e-05, "loss": 1.6954, "step": 3009 }, { "epoch": 0.32529990273424836, "grad_norm": 0.3838809132575989, "learning_rate": 8e-05, "loss": 1.6947, "step": 3010 }, { "epoch": 0.32540797579163516, "grad_norm": 0.3676713705062866, "learning_rate": 8e-05, "loss": 1.7299, "step": 3011 }, { "epoch": 0.32551604884902197, "grad_norm": 0.4285609722137451, "learning_rate": 8e-05, "loss": 1.755, "step": 3012 }, { "epoch": 0.3256241219064087, "grad_norm": 0.43298330903053284, "learning_rate": 8e-05, "loss": 1.7309, "step": 3013 }, { "epoch": 0.3257321949637955, "grad_norm": 0.3864617943763733, "learning_rate": 8e-05, "loss": 1.7414, "step": 3014 }, { "epoch": 0.3258402680211823, "grad_norm": 0.3503204882144928, "learning_rate": 8e-05, "loss": 1.5717, "step": 3015 }, { "epoch": 0.3259483410785691, "grad_norm": 0.4001852571964264, "learning_rate": 8e-05, "loss": 1.7415, "step": 3016 }, { "epoch": 0.32605641413595593, "grad_norm": 0.41210275888442993, "learning_rate": 8e-05, "loss": 1.7325, "step": 3017 }, { "epoch": 0.3261644871933427, "grad_norm": 0.3962157964706421, "learning_rate": 8e-05, "loss": 1.863, "step": 3018 }, { "epoch": 0.3262725602507295, "grad_norm": 0.35352960228919983, "learning_rate": 8e-05, "loss": 1.552, "step": 3019 }, { "epoch": 0.3263806333081163, "grad_norm": 0.3970717787742615, "learning_rate": 8e-05, "loss": 1.6324, "step": 3020 }, { "epoch": 0.3264887063655031, "grad_norm": 0.415105402469635, "learning_rate": 8e-05, "loss": 1.8293, "step": 3021 }, { "epoch": 0.3265967794228899, "grad_norm": 0.4030892848968506, "learning_rate": 8e-05, "loss": 1.8305, "step": 3022 }, { "epoch": 0.32670485248027664, "grad_norm": 0.40116193890571594, "learning_rate": 8e-05, "loss": 1.7658, "step": 3023 }, { "epoch": 0.32681292553766345, "grad_norm": 0.4057353734970093, "learning_rate": 8e-05, "loss": 1.6249, "step": 3024 }, { "epoch": 0.32692099859505025, "grad_norm": 0.3753500282764435, "learning_rate": 8e-05, "loss": 1.6363, "step": 3025 }, { "epoch": 0.32702907165243705, "grad_norm": 0.36307477951049805, "learning_rate": 8e-05, "loss": 1.6871, "step": 3026 }, { "epoch": 0.32713714470982386, "grad_norm": 0.42205509543418884, "learning_rate": 8e-05, "loss": 1.8376, "step": 3027 }, { "epoch": 0.32724521776721066, "grad_norm": 0.38729315996170044, "learning_rate": 8e-05, "loss": 1.8616, "step": 3028 }, { "epoch": 0.3273532908245974, "grad_norm": 0.3738463222980499, "learning_rate": 8e-05, "loss": 1.5999, "step": 3029 }, { "epoch": 0.3274613638819842, "grad_norm": 0.3576565086841583, "learning_rate": 8e-05, "loss": 1.6021, "step": 3030 }, { "epoch": 0.327569436939371, "grad_norm": 0.3857894241809845, "learning_rate": 8e-05, "loss": 1.664, "step": 3031 }, { "epoch": 0.3276775099967578, "grad_norm": 0.37846624851226807, "learning_rate": 8e-05, "loss": 1.6171, "step": 3032 }, { "epoch": 0.3277855830541446, "grad_norm": 0.35405266284942627, "learning_rate": 8e-05, "loss": 1.742, "step": 3033 }, { "epoch": 0.3278936561115314, "grad_norm": 0.38115787506103516, "learning_rate": 8e-05, "loss": 1.7921, "step": 3034 }, { "epoch": 0.3280017291689182, "grad_norm": 0.39969295263290405, "learning_rate": 8e-05, "loss": 1.6825, "step": 3035 }, { "epoch": 0.328109802226305, "grad_norm": 0.4015901982784271, "learning_rate": 8e-05, "loss": 1.852, "step": 3036 }, { "epoch": 0.3282178752836918, "grad_norm": 0.3909691870212555, "learning_rate": 8e-05, "loss": 1.6747, "step": 3037 }, { "epoch": 0.3283259483410786, "grad_norm": 0.4143049716949463, "learning_rate": 8e-05, "loss": 1.7481, "step": 3038 }, { "epoch": 0.32843402139846534, "grad_norm": 0.4304182529449463, "learning_rate": 8e-05, "loss": 1.8353, "step": 3039 }, { "epoch": 0.32854209445585214, "grad_norm": 0.38167572021484375, "learning_rate": 8e-05, "loss": 1.6847, "step": 3040 }, { "epoch": 0.32865016751323894, "grad_norm": 0.3790845274925232, "learning_rate": 8e-05, "loss": 1.7884, "step": 3041 }, { "epoch": 0.32875824057062575, "grad_norm": 0.3853643238544464, "learning_rate": 8e-05, "loss": 1.7687, "step": 3042 }, { "epoch": 0.32886631362801255, "grad_norm": 0.3807254731655121, "learning_rate": 8e-05, "loss": 1.718, "step": 3043 }, { "epoch": 0.32897438668539936, "grad_norm": 0.36925017833709717, "learning_rate": 8e-05, "loss": 1.7273, "step": 3044 }, { "epoch": 0.3290824597427861, "grad_norm": 0.38413169980049133, "learning_rate": 8e-05, "loss": 1.7437, "step": 3045 }, { "epoch": 0.3291905328001729, "grad_norm": 0.395668625831604, "learning_rate": 8e-05, "loss": 1.8495, "step": 3046 }, { "epoch": 0.3292986058575597, "grad_norm": 0.4149981439113617, "learning_rate": 8e-05, "loss": 1.9265, "step": 3047 }, { "epoch": 0.3294066789149465, "grad_norm": 0.38419240713119507, "learning_rate": 8e-05, "loss": 1.8351, "step": 3048 }, { "epoch": 0.3295147519723333, "grad_norm": 0.37794509530067444, "learning_rate": 8e-05, "loss": 1.6931, "step": 3049 }, { "epoch": 0.32962282502972007, "grad_norm": 0.36500272154808044, "learning_rate": 8e-05, "loss": 1.7384, "step": 3050 }, { "epoch": 0.32973089808710687, "grad_norm": 0.36269626021385193, "learning_rate": 8e-05, "loss": 1.697, "step": 3051 }, { "epoch": 0.3298389711444937, "grad_norm": 0.3769221603870392, "learning_rate": 8e-05, "loss": 1.7346, "step": 3052 }, { "epoch": 0.3299470442018805, "grad_norm": 0.39780324697494507, "learning_rate": 8e-05, "loss": 1.7326, "step": 3053 }, { "epoch": 0.3300551172592673, "grad_norm": 0.401928186416626, "learning_rate": 8e-05, "loss": 1.9293, "step": 3054 }, { "epoch": 0.33016319031665403, "grad_norm": 0.4012240171432495, "learning_rate": 8e-05, "loss": 1.694, "step": 3055 }, { "epoch": 0.33027126337404084, "grad_norm": 0.37033188343048096, "learning_rate": 8e-05, "loss": 1.7175, "step": 3056 }, { "epoch": 0.33037933643142764, "grad_norm": 0.43209555745124817, "learning_rate": 8e-05, "loss": 1.9124, "step": 3057 }, { "epoch": 0.33048740948881444, "grad_norm": 0.4278099238872528, "learning_rate": 8e-05, "loss": 1.9179, "step": 3058 }, { "epoch": 0.33059548254620125, "grad_norm": 0.3826291263103485, "learning_rate": 8e-05, "loss": 1.6255, "step": 3059 }, { "epoch": 0.33070355560358805, "grad_norm": 0.376976877450943, "learning_rate": 8e-05, "loss": 1.6399, "step": 3060 }, { "epoch": 0.3308116286609748, "grad_norm": 0.37109488248825073, "learning_rate": 8e-05, "loss": 1.6381, "step": 3061 }, { "epoch": 0.3309197017183616, "grad_norm": 0.4102870225906372, "learning_rate": 8e-05, "loss": 1.8208, "step": 3062 }, { "epoch": 0.3310277747757484, "grad_norm": 0.38958385586738586, "learning_rate": 8e-05, "loss": 1.8929, "step": 3063 }, { "epoch": 0.3311358478331352, "grad_norm": 0.41474154591560364, "learning_rate": 8e-05, "loss": 1.768, "step": 3064 }, { "epoch": 0.331243920890522, "grad_norm": 0.39997225999832153, "learning_rate": 8e-05, "loss": 1.7548, "step": 3065 }, { "epoch": 0.33135199394790876, "grad_norm": 0.3521725535392761, "learning_rate": 8e-05, "loss": 1.6343, "step": 3066 }, { "epoch": 0.33146006700529557, "grad_norm": 0.3650447726249695, "learning_rate": 8e-05, "loss": 1.7048, "step": 3067 }, { "epoch": 0.33156814006268237, "grad_norm": 0.39730072021484375, "learning_rate": 8e-05, "loss": 1.7556, "step": 3068 }, { "epoch": 0.3316762131200692, "grad_norm": 0.37019282579421997, "learning_rate": 8e-05, "loss": 1.6676, "step": 3069 }, { "epoch": 0.331784286177456, "grad_norm": 0.4003468155860901, "learning_rate": 8e-05, "loss": 1.7487, "step": 3070 }, { "epoch": 0.3318923592348427, "grad_norm": 0.3939407169818878, "learning_rate": 8e-05, "loss": 1.76, "step": 3071 }, { "epoch": 0.33200043229222953, "grad_norm": 0.43840292096138, "learning_rate": 8e-05, "loss": 1.9819, "step": 3072 }, { "epoch": 0.33210850534961633, "grad_norm": 0.3808031678199768, "learning_rate": 8e-05, "loss": 1.7616, "step": 3073 }, { "epoch": 0.33221657840700314, "grad_norm": 0.3941723108291626, "learning_rate": 8e-05, "loss": 1.6404, "step": 3074 }, { "epoch": 0.33232465146438994, "grad_norm": 0.36767804622650146, "learning_rate": 8e-05, "loss": 1.6918, "step": 3075 }, { "epoch": 0.33243272452177675, "grad_norm": 0.3735896348953247, "learning_rate": 8e-05, "loss": 1.6804, "step": 3076 }, { "epoch": 0.3325407975791635, "grad_norm": 0.39015597105026245, "learning_rate": 8e-05, "loss": 1.7603, "step": 3077 }, { "epoch": 0.3326488706365503, "grad_norm": 0.44089511036872864, "learning_rate": 8e-05, "loss": 1.9447, "step": 3078 }, { "epoch": 0.3327569436939371, "grad_norm": 0.39315685629844666, "learning_rate": 8e-05, "loss": 1.8167, "step": 3079 }, { "epoch": 0.3328650167513239, "grad_norm": 0.36816471815109253, "learning_rate": 8e-05, "loss": 1.728, "step": 3080 }, { "epoch": 0.3329730898087107, "grad_norm": 0.399775892496109, "learning_rate": 8e-05, "loss": 1.9144, "step": 3081 }, { "epoch": 0.33308116286609746, "grad_norm": 0.3544377088546753, "learning_rate": 8e-05, "loss": 1.6145, "step": 3082 }, { "epoch": 0.33318923592348426, "grad_norm": 0.3989788889884949, "learning_rate": 8e-05, "loss": 1.8408, "step": 3083 }, { "epoch": 0.33329730898087107, "grad_norm": 0.3942987620830536, "learning_rate": 8e-05, "loss": 1.7657, "step": 3084 }, { "epoch": 0.33340538203825787, "grad_norm": 0.35681843757629395, "learning_rate": 8e-05, "loss": 1.6676, "step": 3085 }, { "epoch": 0.3335134550956447, "grad_norm": 0.3745489716529846, "learning_rate": 8e-05, "loss": 1.6656, "step": 3086 }, { "epoch": 0.3336215281530315, "grad_norm": 0.4045359194278717, "learning_rate": 8e-05, "loss": 1.8474, "step": 3087 }, { "epoch": 0.3337296012104182, "grad_norm": 0.3832891583442688, "learning_rate": 8e-05, "loss": 1.7501, "step": 3088 }, { "epoch": 0.33383767426780503, "grad_norm": 0.3704085946083069, "learning_rate": 8e-05, "loss": 1.7004, "step": 3089 }, { "epoch": 0.33394574732519183, "grad_norm": 0.41212600469589233, "learning_rate": 8e-05, "loss": 1.5284, "step": 3090 }, { "epoch": 0.33405382038257864, "grad_norm": 0.3837355077266693, "learning_rate": 8e-05, "loss": 1.7049, "step": 3091 }, { "epoch": 0.33416189343996544, "grad_norm": 0.37648603320121765, "learning_rate": 8e-05, "loss": 1.6555, "step": 3092 }, { "epoch": 0.3342699664973522, "grad_norm": 0.44515129923820496, "learning_rate": 8e-05, "loss": 1.9182, "step": 3093 }, { "epoch": 0.334378039554739, "grad_norm": 0.3692026734352112, "learning_rate": 8e-05, "loss": 1.6226, "step": 3094 }, { "epoch": 0.3344861126121258, "grad_norm": 0.38484832644462585, "learning_rate": 8e-05, "loss": 1.6896, "step": 3095 }, { "epoch": 0.3345941856695126, "grad_norm": 0.3784169852733612, "learning_rate": 8e-05, "loss": 1.5804, "step": 3096 }, { "epoch": 0.3347022587268994, "grad_norm": 0.4062996506690979, "learning_rate": 8e-05, "loss": 1.8089, "step": 3097 }, { "epoch": 0.33481033178428615, "grad_norm": 0.45532190799713135, "learning_rate": 8e-05, "loss": 1.995, "step": 3098 }, { "epoch": 0.33491840484167296, "grad_norm": 0.38555893301963806, "learning_rate": 8e-05, "loss": 1.7749, "step": 3099 }, { "epoch": 0.33502647789905976, "grad_norm": 0.37665918469429016, "learning_rate": 8e-05, "loss": 1.8665, "step": 3100 }, { "epoch": 0.33513455095644656, "grad_norm": 0.3787120580673218, "learning_rate": 8e-05, "loss": 1.7425, "step": 3101 }, { "epoch": 0.33524262401383337, "grad_norm": 0.37780246138572693, "learning_rate": 8e-05, "loss": 1.7697, "step": 3102 }, { "epoch": 0.33535069707122017, "grad_norm": 0.36602702736854553, "learning_rate": 8e-05, "loss": 1.5447, "step": 3103 }, { "epoch": 0.3354587701286069, "grad_norm": 0.3563636243343353, "learning_rate": 8e-05, "loss": 1.6261, "step": 3104 }, { "epoch": 0.3355668431859937, "grad_norm": 0.3499147891998291, "learning_rate": 8e-05, "loss": 1.665, "step": 3105 }, { "epoch": 0.33567491624338053, "grad_norm": 0.3652617633342743, "learning_rate": 8e-05, "loss": 1.7502, "step": 3106 }, { "epoch": 0.33578298930076733, "grad_norm": 0.39372292160987854, "learning_rate": 8e-05, "loss": 1.6593, "step": 3107 }, { "epoch": 0.33589106235815414, "grad_norm": 0.3749449551105499, "learning_rate": 8e-05, "loss": 1.7599, "step": 3108 }, { "epoch": 0.3359991354155409, "grad_norm": 0.4025574028491974, "learning_rate": 8e-05, "loss": 1.7885, "step": 3109 }, { "epoch": 0.3361072084729277, "grad_norm": 0.42658382654190063, "learning_rate": 8e-05, "loss": 1.8234, "step": 3110 }, { "epoch": 0.3362152815303145, "grad_norm": 0.3574577867984772, "learning_rate": 8e-05, "loss": 1.7481, "step": 3111 }, { "epoch": 0.3363233545877013, "grad_norm": 0.41829562187194824, "learning_rate": 8e-05, "loss": 1.7666, "step": 3112 }, { "epoch": 0.3364314276450881, "grad_norm": 0.449555367231369, "learning_rate": 8e-05, "loss": 1.901, "step": 3113 }, { "epoch": 0.33653950070247485, "grad_norm": 0.37352922558784485, "learning_rate": 8e-05, "loss": 1.6738, "step": 3114 }, { "epoch": 0.33664757375986165, "grad_norm": 0.39764904975891113, "learning_rate": 8e-05, "loss": 1.6551, "step": 3115 }, { "epoch": 0.33675564681724846, "grad_norm": 0.3583729565143585, "learning_rate": 8e-05, "loss": 1.7572, "step": 3116 }, { "epoch": 0.33686371987463526, "grad_norm": 0.35891640186309814, "learning_rate": 8e-05, "loss": 1.7714, "step": 3117 }, { "epoch": 0.33697179293202206, "grad_norm": 0.39025938510894775, "learning_rate": 8e-05, "loss": 1.7818, "step": 3118 }, { "epoch": 0.33707986598940887, "grad_norm": 0.3901554048061371, "learning_rate": 8e-05, "loss": 1.7348, "step": 3119 }, { "epoch": 0.3371879390467956, "grad_norm": 0.3619101643562317, "learning_rate": 8e-05, "loss": 1.6234, "step": 3120 }, { "epoch": 0.3372960121041824, "grad_norm": 0.373629629611969, "learning_rate": 8e-05, "loss": 1.6397, "step": 3121 }, { "epoch": 0.3374040851615692, "grad_norm": 0.386385977268219, "learning_rate": 8e-05, "loss": 1.5161, "step": 3122 }, { "epoch": 0.337512158218956, "grad_norm": 0.39593416452407837, "learning_rate": 8e-05, "loss": 1.725, "step": 3123 }, { "epoch": 0.33762023127634283, "grad_norm": 0.44279465079307556, "learning_rate": 8e-05, "loss": 1.7101, "step": 3124 }, { "epoch": 0.3377283043337296, "grad_norm": 0.38848721981048584, "learning_rate": 8e-05, "loss": 1.8649, "step": 3125 }, { "epoch": 0.3378363773911164, "grad_norm": 0.3903926610946655, "learning_rate": 8e-05, "loss": 1.7863, "step": 3126 }, { "epoch": 0.3379444504485032, "grad_norm": 0.3858160674571991, "learning_rate": 8e-05, "loss": 1.776, "step": 3127 }, { "epoch": 0.33805252350589, "grad_norm": 0.39503213763237, "learning_rate": 8e-05, "loss": 1.6768, "step": 3128 }, { "epoch": 0.3381605965632768, "grad_norm": 0.3833896219730377, "learning_rate": 8e-05, "loss": 1.6716, "step": 3129 }, { "epoch": 0.33826866962066354, "grad_norm": 0.3958110213279724, "learning_rate": 8e-05, "loss": 1.747, "step": 3130 }, { "epoch": 0.33837674267805035, "grad_norm": 0.38637807965278625, "learning_rate": 8e-05, "loss": 1.713, "step": 3131 }, { "epoch": 0.33848481573543715, "grad_norm": 0.38812255859375, "learning_rate": 8e-05, "loss": 1.6693, "step": 3132 }, { "epoch": 0.33859288879282395, "grad_norm": 0.38828715682029724, "learning_rate": 8e-05, "loss": 1.7824, "step": 3133 }, { "epoch": 0.33870096185021076, "grad_norm": 0.3985130488872528, "learning_rate": 8e-05, "loss": 1.7277, "step": 3134 }, { "epoch": 0.33880903490759756, "grad_norm": 0.3935237228870392, "learning_rate": 8e-05, "loss": 1.7629, "step": 3135 }, { "epoch": 0.3389171079649843, "grad_norm": 0.38870543241500854, "learning_rate": 8e-05, "loss": 1.7865, "step": 3136 }, { "epoch": 0.3390251810223711, "grad_norm": 0.4428810179233551, "learning_rate": 8e-05, "loss": 1.9752, "step": 3137 }, { "epoch": 0.3391332540797579, "grad_norm": 0.3855208158493042, "learning_rate": 8e-05, "loss": 1.7428, "step": 3138 }, { "epoch": 0.3392413271371447, "grad_norm": 0.3737037479877472, "learning_rate": 8e-05, "loss": 1.8177, "step": 3139 }, { "epoch": 0.3393494001945315, "grad_norm": 0.37476736307144165, "learning_rate": 8e-05, "loss": 1.8322, "step": 3140 }, { "epoch": 0.3394574732519183, "grad_norm": 0.4070238471031189, "learning_rate": 8e-05, "loss": 1.8843, "step": 3141 }, { "epoch": 0.3395655463093051, "grad_norm": 0.37859782576560974, "learning_rate": 8e-05, "loss": 1.7094, "step": 3142 }, { "epoch": 0.3396736193666919, "grad_norm": 0.3812742233276367, "learning_rate": 8e-05, "loss": 1.7236, "step": 3143 }, { "epoch": 0.3397816924240787, "grad_norm": 0.43662509322166443, "learning_rate": 8e-05, "loss": 1.7499, "step": 3144 }, { "epoch": 0.3398897654814655, "grad_norm": 0.38658303022384644, "learning_rate": 8e-05, "loss": 1.7173, "step": 3145 }, { "epoch": 0.33999783853885224, "grad_norm": 0.42068058252334595, "learning_rate": 8e-05, "loss": 1.8062, "step": 3146 }, { "epoch": 0.34010591159623904, "grad_norm": 0.3968547284603119, "learning_rate": 8e-05, "loss": 1.8153, "step": 3147 }, { "epoch": 0.34021398465362584, "grad_norm": 0.3969630002975464, "learning_rate": 8e-05, "loss": 1.8883, "step": 3148 }, { "epoch": 0.34032205771101265, "grad_norm": 0.3536335825920105, "learning_rate": 8e-05, "loss": 1.6495, "step": 3149 }, { "epoch": 0.34043013076839945, "grad_norm": 0.38026586174964905, "learning_rate": 8e-05, "loss": 1.7225, "step": 3150 }, { "epoch": 0.34053820382578626, "grad_norm": 0.3492166996002197, "learning_rate": 8e-05, "loss": 1.6617, "step": 3151 }, { "epoch": 0.340646276883173, "grad_norm": 0.38141706585884094, "learning_rate": 8e-05, "loss": 1.6943, "step": 3152 }, { "epoch": 0.3407543499405598, "grad_norm": 0.3577704429626465, "learning_rate": 8e-05, "loss": 1.625, "step": 3153 }, { "epoch": 0.3408624229979466, "grad_norm": 0.4100503623485565, "learning_rate": 8e-05, "loss": 1.7185, "step": 3154 }, { "epoch": 0.3409704960553334, "grad_norm": 0.40080422163009644, "learning_rate": 8e-05, "loss": 1.6716, "step": 3155 }, { "epoch": 0.3410785691127202, "grad_norm": 0.38559532165527344, "learning_rate": 8e-05, "loss": 1.7915, "step": 3156 }, { "epoch": 0.34118664217010697, "grad_norm": 0.39551371335983276, "learning_rate": 8e-05, "loss": 1.6686, "step": 3157 }, { "epoch": 0.3412947152274938, "grad_norm": 0.38980206847190857, "learning_rate": 8e-05, "loss": 1.7351, "step": 3158 }, { "epoch": 0.3414027882848806, "grad_norm": 0.38582324981689453, "learning_rate": 8e-05, "loss": 1.6744, "step": 3159 }, { "epoch": 0.3415108613422674, "grad_norm": 0.38020479679107666, "learning_rate": 8e-05, "loss": 1.7367, "step": 3160 }, { "epoch": 0.3416189343996542, "grad_norm": 0.3968760371208191, "learning_rate": 8e-05, "loss": 1.7942, "step": 3161 }, { "epoch": 0.34172700745704093, "grad_norm": 0.38403671979904175, "learning_rate": 8e-05, "loss": 1.7275, "step": 3162 }, { "epoch": 0.34183508051442774, "grad_norm": 0.36699190735816956, "learning_rate": 8e-05, "loss": 1.704, "step": 3163 }, { "epoch": 0.34194315357181454, "grad_norm": 0.38546979427337646, "learning_rate": 8e-05, "loss": 1.6797, "step": 3164 }, { "epoch": 0.34205122662920134, "grad_norm": 0.3744828402996063, "learning_rate": 8e-05, "loss": 1.6645, "step": 3165 }, { "epoch": 0.34215929968658815, "grad_norm": 0.38387835025787354, "learning_rate": 8e-05, "loss": 1.8636, "step": 3166 }, { "epoch": 0.34226737274397495, "grad_norm": 0.42923834919929504, "learning_rate": 8e-05, "loss": 1.8915, "step": 3167 }, { "epoch": 0.3423754458013617, "grad_norm": 0.4237186014652252, "learning_rate": 8e-05, "loss": 1.9423, "step": 3168 }, { "epoch": 0.3424835188587485, "grad_norm": 0.39949706196784973, "learning_rate": 8e-05, "loss": 1.7151, "step": 3169 }, { "epoch": 0.3425915919161353, "grad_norm": 0.4231875240802765, "learning_rate": 8e-05, "loss": 1.8786, "step": 3170 }, { "epoch": 0.3426996649735221, "grad_norm": 0.4104372262954712, "learning_rate": 8e-05, "loss": 1.6834, "step": 3171 }, { "epoch": 0.3428077380309089, "grad_norm": 0.3589286506175995, "learning_rate": 8e-05, "loss": 1.5587, "step": 3172 }, { "epoch": 0.34291581108829566, "grad_norm": 0.374222069978714, "learning_rate": 8e-05, "loss": 1.6691, "step": 3173 }, { "epoch": 0.34302388414568247, "grad_norm": 0.39693236351013184, "learning_rate": 8e-05, "loss": 1.8215, "step": 3174 }, { "epoch": 0.34313195720306927, "grad_norm": 0.361100971698761, "learning_rate": 8e-05, "loss": 1.5234, "step": 3175 }, { "epoch": 0.3432400302604561, "grad_norm": 0.47115105390548706, "learning_rate": 8e-05, "loss": 1.9989, "step": 3176 }, { "epoch": 0.3433481033178429, "grad_norm": 0.38971850275993347, "learning_rate": 8e-05, "loss": 1.6616, "step": 3177 }, { "epoch": 0.3434561763752297, "grad_norm": 0.422820508480072, "learning_rate": 8e-05, "loss": 1.8927, "step": 3178 }, { "epoch": 0.34356424943261643, "grad_norm": 0.3851724863052368, "learning_rate": 8e-05, "loss": 1.6103, "step": 3179 }, { "epoch": 0.34367232249000323, "grad_norm": 0.43618062138557434, "learning_rate": 8e-05, "loss": 1.9834, "step": 3180 }, { "epoch": 0.34378039554739004, "grad_norm": 0.39255064725875854, "learning_rate": 8e-05, "loss": 1.8049, "step": 3181 }, { "epoch": 0.34388846860477684, "grad_norm": 0.41304919123649597, "learning_rate": 8e-05, "loss": 1.7901, "step": 3182 }, { "epoch": 0.34399654166216365, "grad_norm": 0.42593175172805786, "learning_rate": 8e-05, "loss": 1.7116, "step": 3183 }, { "epoch": 0.3441046147195504, "grad_norm": 0.39380452036857605, "learning_rate": 8e-05, "loss": 1.8173, "step": 3184 }, { "epoch": 0.3442126877769372, "grad_norm": 0.40321898460388184, "learning_rate": 8e-05, "loss": 1.7012, "step": 3185 }, { "epoch": 0.344320760834324, "grad_norm": 0.35680556297302246, "learning_rate": 8e-05, "loss": 1.7496, "step": 3186 }, { "epoch": 0.3444288338917108, "grad_norm": 0.3606150448322296, "learning_rate": 8e-05, "loss": 1.5744, "step": 3187 }, { "epoch": 0.3445369069490976, "grad_norm": 0.39066699147224426, "learning_rate": 8e-05, "loss": 1.8395, "step": 3188 }, { "epoch": 0.34464498000648436, "grad_norm": 0.37386608123779297, "learning_rate": 8e-05, "loss": 1.7669, "step": 3189 }, { "epoch": 0.34475305306387116, "grad_norm": 0.42777225375175476, "learning_rate": 8e-05, "loss": 1.7297, "step": 3190 }, { "epoch": 0.34486112612125797, "grad_norm": 0.42951536178588867, "learning_rate": 8e-05, "loss": 1.9396, "step": 3191 }, { "epoch": 0.34496919917864477, "grad_norm": 0.3806735575199127, "learning_rate": 8e-05, "loss": 1.6806, "step": 3192 }, { "epoch": 0.3450772722360316, "grad_norm": 0.37916627526283264, "learning_rate": 8e-05, "loss": 1.7228, "step": 3193 }, { "epoch": 0.3451853452934184, "grad_norm": 0.3841780126094818, "learning_rate": 8e-05, "loss": 1.7039, "step": 3194 }, { "epoch": 0.3452934183508051, "grad_norm": 0.3973079025745392, "learning_rate": 8e-05, "loss": 1.6841, "step": 3195 }, { "epoch": 0.34540149140819193, "grad_norm": 0.36865347623825073, "learning_rate": 8e-05, "loss": 1.6411, "step": 3196 }, { "epoch": 0.34550956446557873, "grad_norm": 0.3765410780906677, "learning_rate": 8e-05, "loss": 1.808, "step": 3197 }, { "epoch": 0.34561763752296554, "grad_norm": 0.364969402551651, "learning_rate": 8e-05, "loss": 1.6414, "step": 3198 }, { "epoch": 0.34572571058035234, "grad_norm": 0.4416789710521698, "learning_rate": 8e-05, "loss": 1.8043, "step": 3199 }, { "epoch": 0.3458337836377391, "grad_norm": 0.3804566562175751, "learning_rate": 8e-05, "loss": 1.7166, "step": 3200 }, { "epoch": 0.3459418566951259, "grad_norm": 0.3978241980075836, "learning_rate": 8e-05, "loss": 1.6734, "step": 3201 }, { "epoch": 0.3460499297525127, "grad_norm": 0.4234887361526489, "learning_rate": 8e-05, "loss": 1.8145, "step": 3202 }, { "epoch": 0.3461580028098995, "grad_norm": 0.3662523031234741, "learning_rate": 8e-05, "loss": 1.5956, "step": 3203 }, { "epoch": 0.3462660758672863, "grad_norm": 0.3779642879962921, "learning_rate": 8e-05, "loss": 1.6636, "step": 3204 }, { "epoch": 0.34637414892467305, "grad_norm": 0.38254043459892273, "learning_rate": 8e-05, "loss": 1.7147, "step": 3205 }, { "epoch": 0.34648222198205986, "grad_norm": 0.36932653188705444, "learning_rate": 8e-05, "loss": 1.7334, "step": 3206 }, { "epoch": 0.34659029503944666, "grad_norm": 0.38608768582344055, "learning_rate": 8e-05, "loss": 1.7414, "step": 3207 }, { "epoch": 0.34669836809683346, "grad_norm": 0.39628902077674866, "learning_rate": 8e-05, "loss": 1.748, "step": 3208 }, { "epoch": 0.34680644115422027, "grad_norm": 0.4145163893699646, "learning_rate": 8e-05, "loss": 1.8381, "step": 3209 }, { "epoch": 0.3469145142116071, "grad_norm": 0.3667445480823517, "learning_rate": 8e-05, "loss": 1.7443, "step": 3210 }, { "epoch": 0.3470225872689938, "grad_norm": 0.3577861785888672, "learning_rate": 8e-05, "loss": 1.6976, "step": 3211 }, { "epoch": 0.3471306603263806, "grad_norm": 0.417246013879776, "learning_rate": 8e-05, "loss": 1.6942, "step": 3212 }, { "epoch": 0.34723873338376743, "grad_norm": 0.38905069231987, "learning_rate": 8e-05, "loss": 1.754, "step": 3213 }, { "epoch": 0.34734680644115423, "grad_norm": 0.4353747069835663, "learning_rate": 8e-05, "loss": 1.8437, "step": 3214 }, { "epoch": 0.34745487949854104, "grad_norm": 0.404378205537796, "learning_rate": 8e-05, "loss": 1.8159, "step": 3215 }, { "epoch": 0.3475629525559278, "grad_norm": 0.389313280582428, "learning_rate": 8e-05, "loss": 1.8856, "step": 3216 }, { "epoch": 0.3476710256133146, "grad_norm": 0.36213281750679016, "learning_rate": 8e-05, "loss": 1.604, "step": 3217 }, { "epoch": 0.3477790986707014, "grad_norm": 0.4030488133430481, "learning_rate": 8e-05, "loss": 1.7283, "step": 3218 }, { "epoch": 0.3478871717280882, "grad_norm": 0.3838377594947815, "learning_rate": 8e-05, "loss": 1.6782, "step": 3219 }, { "epoch": 0.347995244785475, "grad_norm": 0.373348206281662, "learning_rate": 8e-05, "loss": 1.6191, "step": 3220 }, { "epoch": 0.34810331784286175, "grad_norm": 0.36474865674972534, "learning_rate": 8e-05, "loss": 1.5536, "step": 3221 }, { "epoch": 0.34821139090024855, "grad_norm": 0.4290372133255005, "learning_rate": 8e-05, "loss": 1.8732, "step": 3222 }, { "epoch": 0.34831946395763536, "grad_norm": 0.4214039444923401, "learning_rate": 8e-05, "loss": 1.7579, "step": 3223 }, { "epoch": 0.34842753701502216, "grad_norm": 0.406820684671402, "learning_rate": 8e-05, "loss": 1.6856, "step": 3224 }, { "epoch": 0.34853561007240896, "grad_norm": 0.37152329087257385, "learning_rate": 8e-05, "loss": 1.7875, "step": 3225 }, { "epoch": 0.34864368312979577, "grad_norm": 0.38109225034713745, "learning_rate": 8e-05, "loss": 1.5189, "step": 3226 }, { "epoch": 0.3487517561871825, "grad_norm": 0.3486393690109253, "learning_rate": 8e-05, "loss": 1.4497, "step": 3227 }, { "epoch": 0.3488598292445693, "grad_norm": 0.4016020596027374, "learning_rate": 8e-05, "loss": 1.4721, "step": 3228 }, { "epoch": 0.3489679023019561, "grad_norm": 0.367669016122818, "learning_rate": 8e-05, "loss": 1.5534, "step": 3229 }, { "epoch": 0.3490759753593429, "grad_norm": 0.42080068588256836, "learning_rate": 8e-05, "loss": 1.8932, "step": 3230 }, { "epoch": 0.34918404841672973, "grad_norm": 0.39687415957450867, "learning_rate": 8e-05, "loss": 1.8092, "step": 3231 }, { "epoch": 0.3492921214741165, "grad_norm": 0.36801913380622864, "learning_rate": 8e-05, "loss": 1.7021, "step": 3232 }, { "epoch": 0.3494001945315033, "grad_norm": 0.3881400227546692, "learning_rate": 8e-05, "loss": 1.7115, "step": 3233 }, { "epoch": 0.3495082675888901, "grad_norm": 0.3757289946079254, "learning_rate": 8e-05, "loss": 1.5718, "step": 3234 }, { "epoch": 0.3496163406462769, "grad_norm": 0.4180893003940582, "learning_rate": 8e-05, "loss": 1.8391, "step": 3235 }, { "epoch": 0.3497244137036637, "grad_norm": 0.4260927438735962, "learning_rate": 8e-05, "loss": 1.908, "step": 3236 }, { "epoch": 0.34983248676105044, "grad_norm": 0.3835902512073517, "learning_rate": 8e-05, "loss": 1.5303, "step": 3237 }, { "epoch": 0.34994055981843725, "grad_norm": 0.4430564343929291, "learning_rate": 8e-05, "loss": 1.4502, "step": 3238 }, { "epoch": 0.35004863287582405, "grad_norm": 0.42253732681274414, "learning_rate": 8e-05, "loss": 1.8428, "step": 3239 }, { "epoch": 0.35015670593321085, "grad_norm": 0.3629271686077118, "learning_rate": 8e-05, "loss": 1.8164, "step": 3240 }, { "epoch": 0.35026477899059766, "grad_norm": 0.3824191391468048, "learning_rate": 8e-05, "loss": 1.7892, "step": 3241 }, { "epoch": 0.35037285204798446, "grad_norm": 0.3858236074447632, "learning_rate": 8e-05, "loss": 1.7041, "step": 3242 }, { "epoch": 0.3504809251053712, "grad_norm": 0.3681716322898865, "learning_rate": 8e-05, "loss": 1.542, "step": 3243 }, { "epoch": 0.350588998162758, "grad_norm": 0.4111172556877136, "learning_rate": 8e-05, "loss": 1.6851, "step": 3244 }, { "epoch": 0.3506970712201448, "grad_norm": 0.3882628083229065, "learning_rate": 8e-05, "loss": 1.6281, "step": 3245 }, { "epoch": 0.3508051442775316, "grad_norm": 0.4404919445514679, "learning_rate": 8e-05, "loss": 1.9407, "step": 3246 }, { "epoch": 0.3509132173349184, "grad_norm": 0.38660356402397156, "learning_rate": 8e-05, "loss": 1.748, "step": 3247 }, { "epoch": 0.3510212903923052, "grad_norm": 0.40403205156326294, "learning_rate": 8e-05, "loss": 1.729, "step": 3248 }, { "epoch": 0.351129363449692, "grad_norm": 0.3867197036743164, "learning_rate": 8e-05, "loss": 1.7883, "step": 3249 }, { "epoch": 0.3512374365070788, "grad_norm": 0.3770262598991394, "learning_rate": 8e-05, "loss": 1.7763, "step": 3250 }, { "epoch": 0.3513455095644656, "grad_norm": 0.37439635396003723, "learning_rate": 8e-05, "loss": 1.7065, "step": 3251 }, { "epoch": 0.3514535826218524, "grad_norm": 0.37813103199005127, "learning_rate": 8e-05, "loss": 1.6222, "step": 3252 }, { "epoch": 0.3515616556792392, "grad_norm": 0.39981237053871155, "learning_rate": 8e-05, "loss": 1.7684, "step": 3253 }, { "epoch": 0.35166972873662594, "grad_norm": 0.44958311319351196, "learning_rate": 8e-05, "loss": 1.7941, "step": 3254 }, { "epoch": 0.35177780179401275, "grad_norm": 0.3765053153038025, "learning_rate": 8e-05, "loss": 1.7046, "step": 3255 }, { "epoch": 0.35188587485139955, "grad_norm": 0.38832759857177734, "learning_rate": 8e-05, "loss": 1.8304, "step": 3256 }, { "epoch": 0.35199394790878635, "grad_norm": 0.3968130052089691, "learning_rate": 8e-05, "loss": 1.766, "step": 3257 }, { "epoch": 0.35210202096617316, "grad_norm": 0.4032117426395416, "learning_rate": 8e-05, "loss": 1.7212, "step": 3258 }, { "epoch": 0.3522100940235599, "grad_norm": 0.3913801610469818, "learning_rate": 8e-05, "loss": 1.6491, "step": 3259 }, { "epoch": 0.3523181670809467, "grad_norm": 0.3861391842365265, "learning_rate": 8e-05, "loss": 1.6915, "step": 3260 }, { "epoch": 0.3524262401383335, "grad_norm": 0.427837997674942, "learning_rate": 8e-05, "loss": 1.7505, "step": 3261 }, { "epoch": 0.3525343131957203, "grad_norm": 0.37296706438064575, "learning_rate": 8e-05, "loss": 1.7805, "step": 3262 }, { "epoch": 0.3526423862531071, "grad_norm": 0.395699679851532, "learning_rate": 8e-05, "loss": 1.8406, "step": 3263 }, { "epoch": 0.35275045931049387, "grad_norm": 0.37352868914604187, "learning_rate": 8e-05, "loss": 1.5707, "step": 3264 }, { "epoch": 0.3528585323678807, "grad_norm": 0.38388049602508545, "learning_rate": 8e-05, "loss": 1.582, "step": 3265 }, { "epoch": 0.3529666054252675, "grad_norm": 0.4121200144290924, "learning_rate": 8e-05, "loss": 1.9766, "step": 3266 }, { "epoch": 0.3530746784826543, "grad_norm": 0.37032556533813477, "learning_rate": 8e-05, "loss": 1.7288, "step": 3267 }, { "epoch": 0.3531827515400411, "grad_norm": 0.4014259874820709, "learning_rate": 8e-05, "loss": 1.8153, "step": 3268 }, { "epoch": 0.3532908245974279, "grad_norm": 0.3905584514141083, "learning_rate": 8e-05, "loss": 1.5817, "step": 3269 }, { "epoch": 0.35339889765481464, "grad_norm": 0.3741360902786255, "learning_rate": 8e-05, "loss": 1.7303, "step": 3270 }, { "epoch": 0.35350697071220144, "grad_norm": 0.3886110186576843, "learning_rate": 8e-05, "loss": 1.5914, "step": 3271 }, { "epoch": 0.35361504376958824, "grad_norm": 0.3873473107814789, "learning_rate": 8e-05, "loss": 1.845, "step": 3272 }, { "epoch": 0.35372311682697505, "grad_norm": 0.3783112168312073, "learning_rate": 8e-05, "loss": 1.5259, "step": 3273 }, { "epoch": 0.35383118988436185, "grad_norm": 0.3893345594406128, "learning_rate": 8e-05, "loss": 1.7142, "step": 3274 }, { "epoch": 0.3539392629417486, "grad_norm": 0.3947809934616089, "learning_rate": 8e-05, "loss": 1.8756, "step": 3275 }, { "epoch": 0.3540473359991354, "grad_norm": 0.37941446900367737, "learning_rate": 8e-05, "loss": 1.7372, "step": 3276 }, { "epoch": 0.3541554090565222, "grad_norm": 0.3749096691608429, "learning_rate": 8e-05, "loss": 1.7362, "step": 3277 }, { "epoch": 0.354263482113909, "grad_norm": 0.4113011360168457, "learning_rate": 8e-05, "loss": 1.7751, "step": 3278 }, { "epoch": 0.3543715551712958, "grad_norm": 0.45836612582206726, "learning_rate": 8e-05, "loss": 1.9683, "step": 3279 }, { "epoch": 0.35447962822868256, "grad_norm": 0.3996044397354126, "learning_rate": 8e-05, "loss": 1.7646, "step": 3280 }, { "epoch": 0.35458770128606937, "grad_norm": 0.36189040541648865, "learning_rate": 8e-05, "loss": 1.6296, "step": 3281 }, { "epoch": 0.35469577434345617, "grad_norm": 0.3680158853530884, "learning_rate": 8e-05, "loss": 1.6799, "step": 3282 }, { "epoch": 0.354803847400843, "grad_norm": 0.38524171710014343, "learning_rate": 8e-05, "loss": 1.8471, "step": 3283 }, { "epoch": 0.3549119204582298, "grad_norm": 0.39464250206947327, "learning_rate": 8e-05, "loss": 1.6951, "step": 3284 }, { "epoch": 0.3550199935156166, "grad_norm": 0.4188738465309143, "learning_rate": 8e-05, "loss": 1.8545, "step": 3285 }, { "epoch": 0.35512806657300333, "grad_norm": 0.3865261673927307, "learning_rate": 8e-05, "loss": 1.745, "step": 3286 }, { "epoch": 0.35523613963039014, "grad_norm": 0.3753278851509094, "learning_rate": 8e-05, "loss": 1.7997, "step": 3287 }, { "epoch": 0.35534421268777694, "grad_norm": 0.3853910267353058, "learning_rate": 8e-05, "loss": 1.7828, "step": 3288 }, { "epoch": 0.35545228574516374, "grad_norm": 0.36639976501464844, "learning_rate": 8e-05, "loss": 1.8848, "step": 3289 }, { "epoch": 0.35556035880255055, "grad_norm": 0.3727981448173523, "learning_rate": 8e-05, "loss": 1.8182, "step": 3290 }, { "epoch": 0.3556684318599373, "grad_norm": 0.35549330711364746, "learning_rate": 8e-05, "loss": 1.7375, "step": 3291 }, { "epoch": 0.3557765049173241, "grad_norm": 0.35986465215682983, "learning_rate": 8e-05, "loss": 1.803, "step": 3292 }, { "epoch": 0.3558845779747109, "grad_norm": 0.37182649970054626, "learning_rate": 8e-05, "loss": 1.6322, "step": 3293 }, { "epoch": 0.3559926510320977, "grad_norm": 0.3608284592628479, "learning_rate": 8e-05, "loss": 1.5897, "step": 3294 }, { "epoch": 0.3561007240894845, "grad_norm": 0.3742779791355133, "learning_rate": 8e-05, "loss": 1.6397, "step": 3295 }, { "epoch": 0.35620879714687126, "grad_norm": 0.4225568175315857, "learning_rate": 8e-05, "loss": 1.8162, "step": 3296 }, { "epoch": 0.35631687020425806, "grad_norm": 0.3930049240589142, "learning_rate": 8e-05, "loss": 1.7037, "step": 3297 }, { "epoch": 0.35642494326164487, "grad_norm": 0.37575840950012207, "learning_rate": 8e-05, "loss": 1.562, "step": 3298 }, { "epoch": 0.35653301631903167, "grad_norm": 0.36371055245399475, "learning_rate": 8e-05, "loss": 1.6417, "step": 3299 }, { "epoch": 0.3566410893764185, "grad_norm": 0.40555402636528015, "learning_rate": 8e-05, "loss": 1.7992, "step": 3300 }, { "epoch": 0.3567491624338053, "grad_norm": 0.35167446732521057, "learning_rate": 8e-05, "loss": 1.6957, "step": 3301 }, { "epoch": 0.356857235491192, "grad_norm": 0.40560516715049744, "learning_rate": 8e-05, "loss": 1.7551, "step": 3302 }, { "epoch": 0.35696530854857883, "grad_norm": 0.39252763986587524, "learning_rate": 8e-05, "loss": 1.6976, "step": 3303 }, { "epoch": 0.35707338160596563, "grad_norm": 0.4329412877559662, "learning_rate": 8e-05, "loss": 1.7916, "step": 3304 }, { "epoch": 0.35718145466335244, "grad_norm": 0.35706016421318054, "learning_rate": 8e-05, "loss": 1.6621, "step": 3305 }, { "epoch": 0.35728952772073924, "grad_norm": 0.3676798641681671, "learning_rate": 8e-05, "loss": 1.7407, "step": 3306 }, { "epoch": 0.357397600778126, "grad_norm": 0.3636726438999176, "learning_rate": 8e-05, "loss": 1.7681, "step": 3307 }, { "epoch": 0.3575056738355128, "grad_norm": 0.4288215637207031, "learning_rate": 8e-05, "loss": 1.6917, "step": 3308 }, { "epoch": 0.3576137468928996, "grad_norm": 0.3651565909385681, "learning_rate": 8e-05, "loss": 1.6398, "step": 3309 }, { "epoch": 0.3577218199502864, "grad_norm": 0.3683786392211914, "learning_rate": 8e-05, "loss": 1.7331, "step": 3310 }, { "epoch": 0.3578298930076732, "grad_norm": 0.44735997915267944, "learning_rate": 8e-05, "loss": 1.7999, "step": 3311 }, { "epoch": 0.35793796606505995, "grad_norm": 0.36708107590675354, "learning_rate": 8e-05, "loss": 1.6989, "step": 3312 }, { "epoch": 0.35804603912244676, "grad_norm": 0.37735840678215027, "learning_rate": 8e-05, "loss": 1.7366, "step": 3313 }, { "epoch": 0.35815411217983356, "grad_norm": 0.4291742444038391, "learning_rate": 8e-05, "loss": 1.9229, "step": 3314 }, { "epoch": 0.35826218523722037, "grad_norm": 0.3743467628955841, "learning_rate": 8e-05, "loss": 1.6357, "step": 3315 }, { "epoch": 0.35837025829460717, "grad_norm": 0.40564483404159546, "learning_rate": 8e-05, "loss": 1.6759, "step": 3316 }, { "epoch": 0.358478331351994, "grad_norm": 0.3940487205982208, "learning_rate": 8e-05, "loss": 1.8145, "step": 3317 }, { "epoch": 0.3585864044093807, "grad_norm": 0.39841681718826294, "learning_rate": 8e-05, "loss": 1.7564, "step": 3318 }, { "epoch": 0.3586944774667675, "grad_norm": 0.3964144289493561, "learning_rate": 8e-05, "loss": 1.7935, "step": 3319 }, { "epoch": 0.35880255052415433, "grad_norm": 0.3572365343570709, "learning_rate": 8e-05, "loss": 1.4962, "step": 3320 }, { "epoch": 0.35891062358154113, "grad_norm": 0.39503270387649536, "learning_rate": 8e-05, "loss": 1.7013, "step": 3321 }, { "epoch": 0.35901869663892794, "grad_norm": 0.4461931586265564, "learning_rate": 8e-05, "loss": 1.6896, "step": 3322 }, { "epoch": 0.3591267696963147, "grad_norm": 0.4256609082221985, "learning_rate": 8e-05, "loss": 1.948, "step": 3323 }, { "epoch": 0.3592348427537015, "grad_norm": 0.36972132325172424, "learning_rate": 8e-05, "loss": 1.7372, "step": 3324 }, { "epoch": 0.3593429158110883, "grad_norm": 0.3727315366268158, "learning_rate": 8e-05, "loss": 1.5656, "step": 3325 }, { "epoch": 0.3594509888684751, "grad_norm": 0.3745492696762085, "learning_rate": 8e-05, "loss": 1.8004, "step": 3326 }, { "epoch": 0.3595590619258619, "grad_norm": 0.4263870120048523, "learning_rate": 8e-05, "loss": 1.9016, "step": 3327 }, { "epoch": 0.35966713498324865, "grad_norm": 0.3618635833263397, "learning_rate": 8e-05, "loss": 1.7851, "step": 3328 }, { "epoch": 0.35977520804063545, "grad_norm": 0.38164615631103516, "learning_rate": 8e-05, "loss": 1.6559, "step": 3329 }, { "epoch": 0.35988328109802226, "grad_norm": 0.41151195764541626, "learning_rate": 8e-05, "loss": 1.8747, "step": 3330 }, { "epoch": 0.35999135415540906, "grad_norm": 0.43405309319496155, "learning_rate": 8e-05, "loss": 1.9899, "step": 3331 }, { "epoch": 0.36009942721279586, "grad_norm": 0.39584243297576904, "learning_rate": 8e-05, "loss": 1.7259, "step": 3332 }, { "epoch": 0.36020750027018267, "grad_norm": 0.3623722493648529, "learning_rate": 8e-05, "loss": 1.7052, "step": 3333 }, { "epoch": 0.3603155733275694, "grad_norm": 0.37232616543769836, "learning_rate": 8e-05, "loss": 1.5709, "step": 3334 }, { "epoch": 0.3604236463849562, "grad_norm": 0.3747260868549347, "learning_rate": 8e-05, "loss": 1.676, "step": 3335 }, { "epoch": 0.360531719442343, "grad_norm": 0.3806154429912567, "learning_rate": 8e-05, "loss": 1.7861, "step": 3336 }, { "epoch": 0.3606397924997298, "grad_norm": 0.3754880428314209, "learning_rate": 8e-05, "loss": 1.768, "step": 3337 }, { "epoch": 0.36074786555711663, "grad_norm": 0.42432963848114014, "learning_rate": 8e-05, "loss": 1.8048, "step": 3338 }, { "epoch": 0.3608559386145034, "grad_norm": 0.379656583070755, "learning_rate": 8e-05, "loss": 1.7883, "step": 3339 }, { "epoch": 0.3609640116718902, "grad_norm": 0.38290655612945557, "learning_rate": 8e-05, "loss": 1.7836, "step": 3340 }, { "epoch": 0.361072084729277, "grad_norm": 0.3752005100250244, "learning_rate": 8e-05, "loss": 1.6715, "step": 3341 }, { "epoch": 0.3611801577866638, "grad_norm": 0.39085811376571655, "learning_rate": 8e-05, "loss": 1.6844, "step": 3342 }, { "epoch": 0.3612882308440506, "grad_norm": 0.3770289719104767, "learning_rate": 8e-05, "loss": 1.6067, "step": 3343 }, { "epoch": 0.3613963039014374, "grad_norm": 0.42762476205825806, "learning_rate": 8e-05, "loss": 1.7253, "step": 3344 }, { "epoch": 0.36150437695882415, "grad_norm": 0.40134239196777344, "learning_rate": 8e-05, "loss": 1.8321, "step": 3345 }, { "epoch": 0.36161245001621095, "grad_norm": 0.4194454848766327, "learning_rate": 8e-05, "loss": 1.8929, "step": 3346 }, { "epoch": 0.36172052307359776, "grad_norm": 0.37847232818603516, "learning_rate": 8e-05, "loss": 1.7654, "step": 3347 }, { "epoch": 0.36182859613098456, "grad_norm": 0.3969729542732239, "learning_rate": 8e-05, "loss": 1.9152, "step": 3348 }, { "epoch": 0.36193666918837136, "grad_norm": 0.4069046974182129, "learning_rate": 8e-05, "loss": 1.7256, "step": 3349 }, { "epoch": 0.3620447422457581, "grad_norm": 0.3857124447822571, "learning_rate": 8e-05, "loss": 1.7474, "step": 3350 }, { "epoch": 0.3621528153031449, "grad_norm": 0.40751519799232483, "learning_rate": 8e-05, "loss": 1.7865, "step": 3351 }, { "epoch": 0.3622608883605317, "grad_norm": 0.40383413434028625, "learning_rate": 8e-05, "loss": 1.8423, "step": 3352 }, { "epoch": 0.3623689614179185, "grad_norm": 0.3802962601184845, "learning_rate": 8e-05, "loss": 1.6872, "step": 3353 }, { "epoch": 0.3624770344753053, "grad_norm": 0.4088432490825653, "learning_rate": 8e-05, "loss": 1.691, "step": 3354 }, { "epoch": 0.3625851075326921, "grad_norm": 0.40254339575767517, "learning_rate": 8e-05, "loss": 1.8043, "step": 3355 }, { "epoch": 0.3626931805900789, "grad_norm": 0.43104445934295654, "learning_rate": 8e-05, "loss": 1.9026, "step": 3356 }, { "epoch": 0.3628012536474657, "grad_norm": 0.3573104739189148, "learning_rate": 8e-05, "loss": 1.7529, "step": 3357 }, { "epoch": 0.3629093267048525, "grad_norm": 0.35362887382507324, "learning_rate": 8e-05, "loss": 1.6582, "step": 3358 }, { "epoch": 0.3630173997622393, "grad_norm": 0.36536577343940735, "learning_rate": 8e-05, "loss": 1.6955, "step": 3359 }, { "epoch": 0.3631254728196261, "grad_norm": 0.4039801359176636, "learning_rate": 8e-05, "loss": 1.895, "step": 3360 }, { "epoch": 0.36323354587701284, "grad_norm": 0.40417835116386414, "learning_rate": 8e-05, "loss": 1.7139, "step": 3361 }, { "epoch": 0.36334161893439965, "grad_norm": 0.3676578402519226, "learning_rate": 8e-05, "loss": 1.6438, "step": 3362 }, { "epoch": 0.36344969199178645, "grad_norm": 0.39095285534858704, "learning_rate": 8e-05, "loss": 1.7849, "step": 3363 }, { "epoch": 0.36355776504917325, "grad_norm": 0.400013267993927, "learning_rate": 8e-05, "loss": 1.8103, "step": 3364 }, { "epoch": 0.36366583810656006, "grad_norm": 0.40802791714668274, "learning_rate": 8e-05, "loss": 1.7873, "step": 3365 }, { "epoch": 0.3637739111639468, "grad_norm": 0.384342759847641, "learning_rate": 8e-05, "loss": 1.8472, "step": 3366 }, { "epoch": 0.3638819842213336, "grad_norm": 0.364883691072464, "learning_rate": 8e-05, "loss": 1.5781, "step": 3367 }, { "epoch": 0.3639900572787204, "grad_norm": 0.40965232253074646, "learning_rate": 8e-05, "loss": 1.8218, "step": 3368 }, { "epoch": 0.3640981303361072, "grad_norm": 0.3790096342563629, "learning_rate": 8e-05, "loss": 1.6618, "step": 3369 }, { "epoch": 0.364206203393494, "grad_norm": 0.382994145154953, "learning_rate": 8e-05, "loss": 1.7175, "step": 3370 }, { "epoch": 0.36431427645088077, "grad_norm": 0.3940376043319702, "learning_rate": 8e-05, "loss": 1.7584, "step": 3371 }, { "epoch": 0.3644223495082676, "grad_norm": 0.39777418971061707, "learning_rate": 8e-05, "loss": 1.8583, "step": 3372 }, { "epoch": 0.3645304225656544, "grad_norm": 0.3720598518848419, "learning_rate": 8e-05, "loss": 1.7014, "step": 3373 }, { "epoch": 0.3646384956230412, "grad_norm": 0.35746097564697266, "learning_rate": 8e-05, "loss": 1.6664, "step": 3374 }, { "epoch": 0.364746568680428, "grad_norm": 0.373651921749115, "learning_rate": 8e-05, "loss": 1.7553, "step": 3375 }, { "epoch": 0.3648546417378148, "grad_norm": 0.3849944770336151, "learning_rate": 8e-05, "loss": 1.7099, "step": 3376 }, { "epoch": 0.36496271479520154, "grad_norm": 0.3637785315513611, "learning_rate": 8e-05, "loss": 1.568, "step": 3377 }, { "epoch": 0.36507078785258834, "grad_norm": 0.4173211455345154, "learning_rate": 8e-05, "loss": 1.8799, "step": 3378 }, { "epoch": 0.36517886090997514, "grad_norm": 0.3767125904560089, "learning_rate": 8e-05, "loss": 1.7062, "step": 3379 }, { "epoch": 0.36528693396736195, "grad_norm": 0.389949232339859, "learning_rate": 8e-05, "loss": 1.821, "step": 3380 }, { "epoch": 0.36539500702474875, "grad_norm": 0.3738099932670593, "learning_rate": 8e-05, "loss": 1.811, "step": 3381 }, { "epoch": 0.3655030800821355, "grad_norm": 0.37187492847442627, "learning_rate": 8e-05, "loss": 1.6008, "step": 3382 }, { "epoch": 0.3656111531395223, "grad_norm": 0.39570388197898865, "learning_rate": 8e-05, "loss": 1.7179, "step": 3383 }, { "epoch": 0.3657192261969091, "grad_norm": 0.3885120451450348, "learning_rate": 8e-05, "loss": 1.7692, "step": 3384 }, { "epoch": 0.3658272992542959, "grad_norm": 0.3875708281993866, "learning_rate": 8e-05, "loss": 1.8532, "step": 3385 }, { "epoch": 0.3659353723116827, "grad_norm": 0.38036608695983887, "learning_rate": 8e-05, "loss": 1.7032, "step": 3386 }, { "epoch": 0.36604344536906946, "grad_norm": 0.3747788369655609, "learning_rate": 8e-05, "loss": 1.7064, "step": 3387 }, { "epoch": 0.36615151842645627, "grad_norm": 0.3880467414855957, "learning_rate": 8e-05, "loss": 1.7259, "step": 3388 }, { "epoch": 0.36625959148384307, "grad_norm": 0.4004300832748413, "learning_rate": 8e-05, "loss": 1.8575, "step": 3389 }, { "epoch": 0.3663676645412299, "grad_norm": 0.4061131179332733, "learning_rate": 8e-05, "loss": 1.8773, "step": 3390 }, { "epoch": 0.3664757375986167, "grad_norm": 0.3810555934906006, "learning_rate": 8e-05, "loss": 1.6489, "step": 3391 }, { "epoch": 0.3665838106560035, "grad_norm": 0.39178967475891113, "learning_rate": 8e-05, "loss": 1.8425, "step": 3392 }, { "epoch": 0.36669188371339023, "grad_norm": 0.3860965967178345, "learning_rate": 8e-05, "loss": 1.7477, "step": 3393 }, { "epoch": 0.36679995677077704, "grad_norm": 0.446403831243515, "learning_rate": 8e-05, "loss": 1.5144, "step": 3394 }, { "epoch": 0.36690802982816384, "grad_norm": 0.3604002296924591, "learning_rate": 8e-05, "loss": 1.6053, "step": 3395 }, { "epoch": 0.36701610288555064, "grad_norm": 0.38694584369659424, "learning_rate": 8e-05, "loss": 1.7023, "step": 3396 }, { "epoch": 0.36712417594293745, "grad_norm": 0.40981706976890564, "learning_rate": 8e-05, "loss": 1.7344, "step": 3397 }, { "epoch": 0.3672322490003242, "grad_norm": 0.4680512249469757, "learning_rate": 8e-05, "loss": 1.8071, "step": 3398 }, { "epoch": 0.367340322057711, "grad_norm": 0.3720478415489197, "learning_rate": 8e-05, "loss": 1.7709, "step": 3399 }, { "epoch": 0.3674483951150978, "grad_norm": 0.36001265048980713, "learning_rate": 8e-05, "loss": 1.6042, "step": 3400 }, { "epoch": 0.3675564681724846, "grad_norm": 0.37720987200737, "learning_rate": 8e-05, "loss": 1.5716, "step": 3401 }, { "epoch": 0.3676645412298714, "grad_norm": 0.41126948595046997, "learning_rate": 8e-05, "loss": 1.7191, "step": 3402 }, { "epoch": 0.36777261428725816, "grad_norm": 0.4064192771911621, "learning_rate": 8e-05, "loss": 1.8318, "step": 3403 }, { "epoch": 0.36788068734464496, "grad_norm": 0.3421679139137268, "learning_rate": 8e-05, "loss": 1.5497, "step": 3404 }, { "epoch": 0.36798876040203177, "grad_norm": 0.39555907249450684, "learning_rate": 8e-05, "loss": 1.8547, "step": 3405 }, { "epoch": 0.36809683345941857, "grad_norm": 0.3768436312675476, "learning_rate": 8e-05, "loss": 1.5171, "step": 3406 }, { "epoch": 0.3682049065168054, "grad_norm": 0.45806387066841125, "learning_rate": 8e-05, "loss": 1.7317, "step": 3407 }, { "epoch": 0.3683129795741922, "grad_norm": 0.3954867124557495, "learning_rate": 8e-05, "loss": 1.7231, "step": 3408 }, { "epoch": 0.3684210526315789, "grad_norm": 0.4032396376132965, "learning_rate": 8e-05, "loss": 1.783, "step": 3409 }, { "epoch": 0.36852912568896573, "grad_norm": 0.3578210175037384, "learning_rate": 8e-05, "loss": 1.6938, "step": 3410 }, { "epoch": 0.36863719874635253, "grad_norm": 0.37989288568496704, "learning_rate": 8e-05, "loss": 1.8005, "step": 3411 }, { "epoch": 0.36874527180373934, "grad_norm": 0.42165324091911316, "learning_rate": 8e-05, "loss": 1.4566, "step": 3412 }, { "epoch": 0.36885334486112614, "grad_norm": 0.3587247133255005, "learning_rate": 8e-05, "loss": 1.6294, "step": 3413 }, { "epoch": 0.3689614179185129, "grad_norm": 0.39802420139312744, "learning_rate": 8e-05, "loss": 1.7039, "step": 3414 }, { "epoch": 0.3690694909758997, "grad_norm": 0.4061676263809204, "learning_rate": 8e-05, "loss": 1.7665, "step": 3415 }, { "epoch": 0.3691775640332865, "grad_norm": 0.4068739116191864, "learning_rate": 8e-05, "loss": 1.6524, "step": 3416 }, { "epoch": 0.3692856370906733, "grad_norm": 0.44787344336509705, "learning_rate": 8e-05, "loss": 1.7114, "step": 3417 }, { "epoch": 0.3693937101480601, "grad_norm": 0.3850559890270233, "learning_rate": 8e-05, "loss": 1.8842, "step": 3418 }, { "epoch": 0.3695017832054469, "grad_norm": 0.38400906324386597, "learning_rate": 8e-05, "loss": 1.6283, "step": 3419 }, { "epoch": 0.36960985626283366, "grad_norm": 0.4191437065601349, "learning_rate": 8e-05, "loss": 2.0712, "step": 3420 }, { "epoch": 0.36971792932022046, "grad_norm": 0.37882089614868164, "learning_rate": 8e-05, "loss": 1.6564, "step": 3421 }, { "epoch": 0.36982600237760727, "grad_norm": 0.37706458568573, "learning_rate": 8e-05, "loss": 1.8448, "step": 3422 }, { "epoch": 0.36993407543499407, "grad_norm": 0.3884028196334839, "learning_rate": 8e-05, "loss": 1.7093, "step": 3423 }, { "epoch": 0.3700421484923809, "grad_norm": 0.3944375813007355, "learning_rate": 8e-05, "loss": 1.7908, "step": 3424 }, { "epoch": 0.3701502215497676, "grad_norm": 0.37425515055656433, "learning_rate": 8e-05, "loss": 1.7829, "step": 3425 }, { "epoch": 0.3702582946071544, "grad_norm": 0.372321754693985, "learning_rate": 8e-05, "loss": 1.741, "step": 3426 }, { "epoch": 0.37036636766454123, "grad_norm": 0.40048331022262573, "learning_rate": 8e-05, "loss": 1.6186, "step": 3427 }, { "epoch": 0.37047444072192803, "grad_norm": 0.38679149746894836, "learning_rate": 8e-05, "loss": 1.7213, "step": 3428 }, { "epoch": 0.37058251377931484, "grad_norm": 0.40259772539138794, "learning_rate": 8e-05, "loss": 1.7207, "step": 3429 }, { "epoch": 0.3706905868367016, "grad_norm": 0.40421050786972046, "learning_rate": 8e-05, "loss": 1.7102, "step": 3430 }, { "epoch": 0.3707986598940884, "grad_norm": 0.36750131845474243, "learning_rate": 8e-05, "loss": 1.7066, "step": 3431 }, { "epoch": 0.3709067329514752, "grad_norm": 0.4113597571849823, "learning_rate": 8e-05, "loss": 1.7967, "step": 3432 }, { "epoch": 0.371014806008862, "grad_norm": 0.3771274983882904, "learning_rate": 8e-05, "loss": 1.7778, "step": 3433 }, { "epoch": 0.3711228790662488, "grad_norm": 0.44363629817962646, "learning_rate": 8e-05, "loss": 1.7805, "step": 3434 }, { "epoch": 0.3712309521236356, "grad_norm": 0.36660289764404297, "learning_rate": 8e-05, "loss": 1.6355, "step": 3435 }, { "epoch": 0.37133902518102235, "grad_norm": 0.41452327370643616, "learning_rate": 8e-05, "loss": 1.9428, "step": 3436 }, { "epoch": 0.37144709823840916, "grad_norm": 0.4000818431377411, "learning_rate": 8e-05, "loss": 1.7167, "step": 3437 }, { "epoch": 0.37155517129579596, "grad_norm": 0.42190349102020264, "learning_rate": 8e-05, "loss": 1.847, "step": 3438 }, { "epoch": 0.37166324435318276, "grad_norm": 0.3889029622077942, "learning_rate": 8e-05, "loss": 1.7787, "step": 3439 }, { "epoch": 0.37177131741056957, "grad_norm": 0.3782348334789276, "learning_rate": 8e-05, "loss": 1.6333, "step": 3440 }, { "epoch": 0.3718793904679563, "grad_norm": 0.396033376455307, "learning_rate": 8e-05, "loss": 1.6404, "step": 3441 }, { "epoch": 0.3719874635253431, "grad_norm": 0.44668343663215637, "learning_rate": 8e-05, "loss": 1.9855, "step": 3442 }, { "epoch": 0.3720955365827299, "grad_norm": 0.38046517968177795, "learning_rate": 8e-05, "loss": 1.665, "step": 3443 }, { "epoch": 0.37220360964011673, "grad_norm": 0.35904237627983093, "learning_rate": 8e-05, "loss": 1.5858, "step": 3444 }, { "epoch": 0.37231168269750353, "grad_norm": 0.41577133536338806, "learning_rate": 8e-05, "loss": 1.7926, "step": 3445 }, { "epoch": 0.3724197557548903, "grad_norm": 0.3733823895454407, "learning_rate": 8e-05, "loss": 1.5635, "step": 3446 }, { "epoch": 0.3725278288122771, "grad_norm": 0.391984224319458, "learning_rate": 8e-05, "loss": 1.837, "step": 3447 }, { "epoch": 0.3726359018696639, "grad_norm": 0.39847758412361145, "learning_rate": 8e-05, "loss": 1.7879, "step": 3448 }, { "epoch": 0.3727439749270507, "grad_norm": 0.38917872309684753, "learning_rate": 8e-05, "loss": 1.7321, "step": 3449 }, { "epoch": 0.3728520479844375, "grad_norm": 0.38308292627334595, "learning_rate": 8e-05, "loss": 1.7025, "step": 3450 }, { "epoch": 0.3729601210418243, "grad_norm": 0.3901844918727875, "learning_rate": 8e-05, "loss": 1.8088, "step": 3451 }, { "epoch": 0.37306819409921105, "grad_norm": 0.3926748037338257, "learning_rate": 8e-05, "loss": 1.704, "step": 3452 }, { "epoch": 0.37317626715659785, "grad_norm": 0.41543129086494446, "learning_rate": 8e-05, "loss": 1.9184, "step": 3453 }, { "epoch": 0.37328434021398466, "grad_norm": 0.3870210349559784, "learning_rate": 8e-05, "loss": 1.657, "step": 3454 }, { "epoch": 0.37339241327137146, "grad_norm": 0.42889267206192017, "learning_rate": 8e-05, "loss": 1.8621, "step": 3455 }, { "epoch": 0.37350048632875826, "grad_norm": 0.39543601870536804, "learning_rate": 8e-05, "loss": 1.7097, "step": 3456 }, { "epoch": 0.373608559386145, "grad_norm": 0.3894956409931183, "learning_rate": 8e-05, "loss": 1.6513, "step": 3457 }, { "epoch": 0.3737166324435318, "grad_norm": 0.4336859881877899, "learning_rate": 8e-05, "loss": 1.871, "step": 3458 }, { "epoch": 0.3738247055009186, "grad_norm": 0.3818238079547882, "learning_rate": 8e-05, "loss": 1.7545, "step": 3459 }, { "epoch": 0.3739327785583054, "grad_norm": 0.365071564912796, "learning_rate": 8e-05, "loss": 1.6246, "step": 3460 }, { "epoch": 0.3740408516156922, "grad_norm": 0.4249206781387329, "learning_rate": 8e-05, "loss": 1.7509, "step": 3461 }, { "epoch": 0.374148924673079, "grad_norm": 0.37332838773727417, "learning_rate": 8e-05, "loss": 1.8074, "step": 3462 }, { "epoch": 0.3742569977304658, "grad_norm": 0.38897836208343506, "learning_rate": 8e-05, "loss": 1.7653, "step": 3463 }, { "epoch": 0.3743650707878526, "grad_norm": 0.35265931487083435, "learning_rate": 8e-05, "loss": 1.5697, "step": 3464 }, { "epoch": 0.3744731438452394, "grad_norm": 0.4293357729911804, "learning_rate": 8e-05, "loss": 1.9529, "step": 3465 }, { "epoch": 0.3745812169026262, "grad_norm": 0.4293860197067261, "learning_rate": 8e-05, "loss": 1.7133, "step": 3466 }, { "epoch": 0.374689289960013, "grad_norm": 0.39301517605781555, "learning_rate": 8e-05, "loss": 1.8131, "step": 3467 }, { "epoch": 0.37479736301739974, "grad_norm": 0.3787461817264557, "learning_rate": 8e-05, "loss": 1.7794, "step": 3468 }, { "epoch": 0.37490543607478655, "grad_norm": 0.38320720195770264, "learning_rate": 8e-05, "loss": 1.7574, "step": 3469 }, { "epoch": 0.37501350913217335, "grad_norm": 0.4076857566833496, "learning_rate": 8e-05, "loss": 1.733, "step": 3470 }, { "epoch": 0.37512158218956015, "grad_norm": 0.4291606545448303, "learning_rate": 8e-05, "loss": 1.7554, "step": 3471 }, { "epoch": 0.37522965524694696, "grad_norm": 0.38234496116638184, "learning_rate": 8e-05, "loss": 1.759, "step": 3472 }, { "epoch": 0.3753377283043337, "grad_norm": 0.36409610509872437, "learning_rate": 8e-05, "loss": 1.6945, "step": 3473 }, { "epoch": 0.3754458013617205, "grad_norm": 0.39078328013420105, "learning_rate": 8e-05, "loss": 1.756, "step": 3474 }, { "epoch": 0.3755538744191073, "grad_norm": 0.41482365131378174, "learning_rate": 8e-05, "loss": 1.7691, "step": 3475 }, { "epoch": 0.3756619474764941, "grad_norm": 0.3957419693470001, "learning_rate": 8e-05, "loss": 1.7109, "step": 3476 }, { "epoch": 0.3757700205338809, "grad_norm": 0.39946168661117554, "learning_rate": 8e-05, "loss": 1.7913, "step": 3477 }, { "epoch": 0.37587809359126767, "grad_norm": 0.3596000671386719, "learning_rate": 8e-05, "loss": 1.6198, "step": 3478 }, { "epoch": 0.3759861666486545, "grad_norm": 0.40265509486198425, "learning_rate": 8e-05, "loss": 1.6563, "step": 3479 }, { "epoch": 0.3760942397060413, "grad_norm": 0.3781740367412567, "learning_rate": 8e-05, "loss": 1.756, "step": 3480 }, { "epoch": 0.3762023127634281, "grad_norm": 0.39202919602394104, "learning_rate": 8e-05, "loss": 1.6537, "step": 3481 }, { "epoch": 0.3763103858208149, "grad_norm": 0.3824671804904938, "learning_rate": 8e-05, "loss": 1.6935, "step": 3482 }, { "epoch": 0.3764184588782017, "grad_norm": 0.4187178313732147, "learning_rate": 8e-05, "loss": 1.8507, "step": 3483 }, { "epoch": 0.37652653193558844, "grad_norm": 0.3949556350708008, "learning_rate": 8e-05, "loss": 1.689, "step": 3484 }, { "epoch": 0.37663460499297524, "grad_norm": 0.4136996865272522, "learning_rate": 8e-05, "loss": 1.5416, "step": 3485 }, { "epoch": 0.37674267805036205, "grad_norm": 0.4144202172756195, "learning_rate": 8e-05, "loss": 1.8402, "step": 3486 }, { "epoch": 0.37685075110774885, "grad_norm": 0.38980695605278015, "learning_rate": 8e-05, "loss": 1.6762, "step": 3487 }, { "epoch": 0.37695882416513565, "grad_norm": 0.3771219551563263, "learning_rate": 8e-05, "loss": 1.6669, "step": 3488 }, { "epoch": 0.3770668972225224, "grad_norm": 0.3977028429508209, "learning_rate": 8e-05, "loss": 1.6849, "step": 3489 }, { "epoch": 0.3771749702799092, "grad_norm": 0.3851807117462158, "learning_rate": 8e-05, "loss": 1.6491, "step": 3490 }, { "epoch": 0.377283043337296, "grad_norm": 0.39543524384498596, "learning_rate": 8e-05, "loss": 1.6327, "step": 3491 }, { "epoch": 0.3773911163946828, "grad_norm": 0.4074676036834717, "learning_rate": 8e-05, "loss": 1.703, "step": 3492 }, { "epoch": 0.3774991894520696, "grad_norm": 0.4018513262271881, "learning_rate": 8e-05, "loss": 1.7717, "step": 3493 }, { "epoch": 0.37760726250945637, "grad_norm": 0.39017102122306824, "learning_rate": 8e-05, "loss": 1.7066, "step": 3494 }, { "epoch": 0.37771533556684317, "grad_norm": 0.38313257694244385, "learning_rate": 8e-05, "loss": 1.737, "step": 3495 }, { "epoch": 0.37782340862423, "grad_norm": 0.393898606300354, "learning_rate": 8e-05, "loss": 1.6741, "step": 3496 }, { "epoch": 0.3779314816816168, "grad_norm": 0.37563177943229675, "learning_rate": 8e-05, "loss": 1.7086, "step": 3497 }, { "epoch": 0.3780395547390036, "grad_norm": 0.3859784007072449, "learning_rate": 8e-05, "loss": 1.6999, "step": 3498 }, { "epoch": 0.3781476277963904, "grad_norm": 0.40964189171791077, "learning_rate": 8e-05, "loss": 1.8494, "step": 3499 }, { "epoch": 0.37825570085377713, "grad_norm": 0.37586721777915955, "learning_rate": 8e-05, "loss": 1.696, "step": 3500 }, { "epoch": 0.37836377391116394, "grad_norm": 0.41263341903686523, "learning_rate": 8e-05, "loss": 1.8456, "step": 3501 }, { "epoch": 0.37847184696855074, "grad_norm": 0.42650192975997925, "learning_rate": 8e-05, "loss": 1.8379, "step": 3502 }, { "epoch": 0.37857992002593754, "grad_norm": 0.38107308745384216, "learning_rate": 8e-05, "loss": 1.7373, "step": 3503 }, { "epoch": 0.37868799308332435, "grad_norm": 0.4120958745479584, "learning_rate": 8e-05, "loss": 1.7522, "step": 3504 }, { "epoch": 0.3787960661407111, "grad_norm": 0.40307390689849854, "learning_rate": 8e-05, "loss": 1.6929, "step": 3505 }, { "epoch": 0.3789041391980979, "grad_norm": 0.37662678956985474, "learning_rate": 8e-05, "loss": 1.6681, "step": 3506 }, { "epoch": 0.3790122122554847, "grad_norm": 0.38320988416671753, "learning_rate": 8e-05, "loss": 1.6575, "step": 3507 }, { "epoch": 0.3791202853128715, "grad_norm": 0.36323174834251404, "learning_rate": 8e-05, "loss": 1.5795, "step": 3508 }, { "epoch": 0.3792283583702583, "grad_norm": 0.37993577122688293, "learning_rate": 8e-05, "loss": 1.7888, "step": 3509 }, { "epoch": 0.3793364314276451, "grad_norm": 0.3733411729335785, "learning_rate": 8e-05, "loss": 1.7635, "step": 3510 }, { "epoch": 0.37944450448503186, "grad_norm": 0.36715659499168396, "learning_rate": 8e-05, "loss": 1.6497, "step": 3511 }, { "epoch": 0.37955257754241867, "grad_norm": 0.4003114700317383, "learning_rate": 8e-05, "loss": 1.8552, "step": 3512 }, { "epoch": 0.37966065059980547, "grad_norm": 0.3740144371986389, "learning_rate": 8e-05, "loss": 1.6465, "step": 3513 }, { "epoch": 0.3797687236571923, "grad_norm": 0.3640976548194885, "learning_rate": 8e-05, "loss": 1.6542, "step": 3514 }, { "epoch": 0.3798767967145791, "grad_norm": 0.3802964389324188, "learning_rate": 8e-05, "loss": 1.7625, "step": 3515 }, { "epoch": 0.3799848697719658, "grad_norm": 0.38248687982559204, "learning_rate": 8e-05, "loss": 1.6763, "step": 3516 }, { "epoch": 0.38009294282935263, "grad_norm": 0.41036197543144226, "learning_rate": 8e-05, "loss": 1.8767, "step": 3517 }, { "epoch": 0.38020101588673944, "grad_norm": 0.363010048866272, "learning_rate": 8e-05, "loss": 1.6469, "step": 3518 }, { "epoch": 0.38030908894412624, "grad_norm": 0.3722565472126007, "learning_rate": 8e-05, "loss": 1.7319, "step": 3519 }, { "epoch": 0.38041716200151304, "grad_norm": 0.40870213508605957, "learning_rate": 8e-05, "loss": 1.7385, "step": 3520 }, { "epoch": 0.3805252350588998, "grad_norm": 0.4093620479106903, "learning_rate": 8e-05, "loss": 1.728, "step": 3521 }, { "epoch": 0.3806333081162866, "grad_norm": 0.431355357170105, "learning_rate": 8e-05, "loss": 1.9047, "step": 3522 }, { "epoch": 0.3807413811736734, "grad_norm": 0.39238104224205017, "learning_rate": 8e-05, "loss": 1.7293, "step": 3523 }, { "epoch": 0.3808494542310602, "grad_norm": 0.4006044566631317, "learning_rate": 8e-05, "loss": 1.8088, "step": 3524 }, { "epoch": 0.380957527288447, "grad_norm": 0.3569253981113434, "learning_rate": 8e-05, "loss": 1.7129, "step": 3525 }, { "epoch": 0.3810656003458338, "grad_norm": 0.41473475098609924, "learning_rate": 8e-05, "loss": 1.8542, "step": 3526 }, { "epoch": 0.38117367340322056, "grad_norm": 0.36449646949768066, "learning_rate": 8e-05, "loss": 1.5846, "step": 3527 }, { "epoch": 0.38128174646060736, "grad_norm": 0.38516658544540405, "learning_rate": 8e-05, "loss": 1.785, "step": 3528 }, { "epoch": 0.38138981951799417, "grad_norm": 0.40685662627220154, "learning_rate": 8e-05, "loss": 1.8684, "step": 3529 }, { "epoch": 0.38149789257538097, "grad_norm": 0.416901171207428, "learning_rate": 8e-05, "loss": 1.782, "step": 3530 }, { "epoch": 0.3816059656327678, "grad_norm": 0.37192219495773315, "learning_rate": 8e-05, "loss": 1.7765, "step": 3531 }, { "epoch": 0.3817140386901545, "grad_norm": 0.38957494497299194, "learning_rate": 8e-05, "loss": 1.7564, "step": 3532 }, { "epoch": 0.3818221117475413, "grad_norm": 0.3893946409225464, "learning_rate": 8e-05, "loss": 1.7554, "step": 3533 }, { "epoch": 0.38193018480492813, "grad_norm": 0.42309218645095825, "learning_rate": 8e-05, "loss": 1.7674, "step": 3534 }, { "epoch": 0.38203825786231493, "grad_norm": 0.4150150418281555, "learning_rate": 8e-05, "loss": 1.6225, "step": 3535 }, { "epoch": 0.38214633091970174, "grad_norm": 0.42698100209236145, "learning_rate": 8e-05, "loss": 1.7837, "step": 3536 }, { "epoch": 0.3822544039770885, "grad_norm": 0.3650544285774231, "learning_rate": 8e-05, "loss": 1.7287, "step": 3537 }, { "epoch": 0.3823624770344753, "grad_norm": 0.38779762387275696, "learning_rate": 8e-05, "loss": 1.8059, "step": 3538 }, { "epoch": 0.3824705500918621, "grad_norm": 0.37549030780792236, "learning_rate": 8e-05, "loss": 1.7605, "step": 3539 }, { "epoch": 0.3825786231492489, "grad_norm": 0.38927751779556274, "learning_rate": 8e-05, "loss": 1.6573, "step": 3540 }, { "epoch": 0.3826866962066357, "grad_norm": 0.3979440927505493, "learning_rate": 8e-05, "loss": 1.7416, "step": 3541 }, { "epoch": 0.3827947692640225, "grad_norm": 0.4279042184352875, "learning_rate": 8e-05, "loss": 1.7218, "step": 3542 }, { "epoch": 0.38290284232140925, "grad_norm": 0.37709760665893555, "learning_rate": 8e-05, "loss": 1.7152, "step": 3543 }, { "epoch": 0.38301091537879606, "grad_norm": 0.41234949231147766, "learning_rate": 8e-05, "loss": 1.6119, "step": 3544 }, { "epoch": 0.38311898843618286, "grad_norm": 0.37349435687065125, "learning_rate": 8e-05, "loss": 1.6874, "step": 3545 }, { "epoch": 0.38322706149356967, "grad_norm": 0.41712650656700134, "learning_rate": 8e-05, "loss": 1.8228, "step": 3546 }, { "epoch": 0.38333513455095647, "grad_norm": 0.4010965824127197, "learning_rate": 8e-05, "loss": 1.592, "step": 3547 }, { "epoch": 0.3834432076083432, "grad_norm": 0.3669918477535248, "learning_rate": 8e-05, "loss": 1.6656, "step": 3548 }, { "epoch": 0.38355128066573, "grad_norm": 0.40519875288009644, "learning_rate": 8e-05, "loss": 1.8051, "step": 3549 }, { "epoch": 0.3836593537231168, "grad_norm": 0.3981151878833771, "learning_rate": 8e-05, "loss": 1.8568, "step": 3550 }, { "epoch": 0.38376742678050363, "grad_norm": 0.3808800280094147, "learning_rate": 8e-05, "loss": 1.7777, "step": 3551 }, { "epoch": 0.38387549983789043, "grad_norm": 0.3973964750766754, "learning_rate": 8e-05, "loss": 1.8324, "step": 3552 }, { "epoch": 0.3839835728952772, "grad_norm": 0.37368136644363403, "learning_rate": 8e-05, "loss": 1.814, "step": 3553 }, { "epoch": 0.384091645952664, "grad_norm": 0.3900277316570282, "learning_rate": 8e-05, "loss": 1.8975, "step": 3554 }, { "epoch": 0.3841997190100508, "grad_norm": 0.4191446006298065, "learning_rate": 8e-05, "loss": 1.8391, "step": 3555 }, { "epoch": 0.3843077920674376, "grad_norm": 0.44933372735977173, "learning_rate": 8e-05, "loss": 1.9008, "step": 3556 }, { "epoch": 0.3844158651248244, "grad_norm": 0.44189462065696716, "learning_rate": 8e-05, "loss": 1.9726, "step": 3557 }, { "epoch": 0.3845239381822112, "grad_norm": 0.37624919414520264, "learning_rate": 8e-05, "loss": 1.6757, "step": 3558 }, { "epoch": 0.38463201123959795, "grad_norm": 0.3720269501209259, "learning_rate": 8e-05, "loss": 1.682, "step": 3559 }, { "epoch": 0.38474008429698475, "grad_norm": 0.4034567177295685, "learning_rate": 8e-05, "loss": 1.7408, "step": 3560 }, { "epoch": 0.38484815735437156, "grad_norm": 0.3816892206668854, "learning_rate": 8e-05, "loss": 1.7114, "step": 3561 }, { "epoch": 0.38495623041175836, "grad_norm": 0.404674768447876, "learning_rate": 8e-05, "loss": 1.7227, "step": 3562 }, { "epoch": 0.38506430346914516, "grad_norm": 0.44607359170913696, "learning_rate": 8e-05, "loss": 1.8034, "step": 3563 }, { "epoch": 0.3851723765265319, "grad_norm": 0.3940983712673187, "learning_rate": 8e-05, "loss": 1.8253, "step": 3564 }, { "epoch": 0.3852804495839187, "grad_norm": 0.40224984288215637, "learning_rate": 8e-05, "loss": 1.801, "step": 3565 }, { "epoch": 0.3853885226413055, "grad_norm": 0.39375922083854675, "learning_rate": 8e-05, "loss": 1.7587, "step": 3566 }, { "epoch": 0.3854965956986923, "grad_norm": 0.403225839138031, "learning_rate": 8e-05, "loss": 1.779, "step": 3567 }, { "epoch": 0.3856046687560791, "grad_norm": 0.4228249788284302, "learning_rate": 8e-05, "loss": 1.7979, "step": 3568 }, { "epoch": 0.3857127418134659, "grad_norm": 0.39948219060897827, "learning_rate": 8e-05, "loss": 1.8201, "step": 3569 }, { "epoch": 0.3858208148708527, "grad_norm": 0.3792552947998047, "learning_rate": 8e-05, "loss": 1.7587, "step": 3570 }, { "epoch": 0.3859288879282395, "grad_norm": 0.3772358298301697, "learning_rate": 8e-05, "loss": 1.5684, "step": 3571 }, { "epoch": 0.3860369609856263, "grad_norm": 0.3831201493740082, "learning_rate": 8e-05, "loss": 1.7019, "step": 3572 }, { "epoch": 0.3861450340430131, "grad_norm": 0.419120728969574, "learning_rate": 8e-05, "loss": 1.7269, "step": 3573 }, { "epoch": 0.3862531071003999, "grad_norm": 0.3983045518398285, "learning_rate": 8e-05, "loss": 1.7789, "step": 3574 }, { "epoch": 0.38636118015778664, "grad_norm": 0.40484148263931274, "learning_rate": 8e-05, "loss": 1.6836, "step": 3575 }, { "epoch": 0.38646925321517345, "grad_norm": 0.4391734004020691, "learning_rate": 8e-05, "loss": 1.6993, "step": 3576 }, { "epoch": 0.38657732627256025, "grad_norm": 0.4135514199733734, "learning_rate": 8e-05, "loss": 1.9047, "step": 3577 }, { "epoch": 0.38668539932994705, "grad_norm": 0.38191351294517517, "learning_rate": 8e-05, "loss": 1.4872, "step": 3578 }, { "epoch": 0.38679347238733386, "grad_norm": 0.4104292094707489, "learning_rate": 8e-05, "loss": 1.7859, "step": 3579 }, { "epoch": 0.3869015454447206, "grad_norm": 0.39411255717277527, "learning_rate": 8e-05, "loss": 1.7829, "step": 3580 }, { "epoch": 0.3870096185021074, "grad_norm": 0.360411137342453, "learning_rate": 8e-05, "loss": 1.5311, "step": 3581 }, { "epoch": 0.3871176915594942, "grad_norm": 0.43718859553337097, "learning_rate": 8e-05, "loss": 1.6604, "step": 3582 }, { "epoch": 0.387225764616881, "grad_norm": 0.38726693391799927, "learning_rate": 8e-05, "loss": 1.758, "step": 3583 }, { "epoch": 0.3873338376742678, "grad_norm": 0.3864075541496277, "learning_rate": 8e-05, "loss": 1.7241, "step": 3584 }, { "epoch": 0.3874419107316546, "grad_norm": 0.40262165665626526, "learning_rate": 8e-05, "loss": 1.6771, "step": 3585 }, { "epoch": 0.3875499837890414, "grad_norm": 0.37532612681388855, "learning_rate": 8e-05, "loss": 1.6484, "step": 3586 }, { "epoch": 0.3876580568464282, "grad_norm": 0.38790810108184814, "learning_rate": 8e-05, "loss": 1.8006, "step": 3587 }, { "epoch": 0.387766129903815, "grad_norm": 0.3883034884929657, "learning_rate": 8e-05, "loss": 1.7368, "step": 3588 }, { "epoch": 0.3878742029612018, "grad_norm": 0.4685097336769104, "learning_rate": 8e-05, "loss": 1.6946, "step": 3589 }, { "epoch": 0.3879822760185886, "grad_norm": 0.4104948043823242, "learning_rate": 8e-05, "loss": 1.7799, "step": 3590 }, { "epoch": 0.38809034907597534, "grad_norm": 0.4495389759540558, "learning_rate": 8e-05, "loss": 1.715, "step": 3591 }, { "epoch": 0.38819842213336214, "grad_norm": 0.4029026925563812, "learning_rate": 8e-05, "loss": 1.7304, "step": 3592 }, { "epoch": 0.38830649519074895, "grad_norm": 0.3995441794395447, "learning_rate": 8e-05, "loss": 1.698, "step": 3593 }, { "epoch": 0.38841456824813575, "grad_norm": 0.4161490201950073, "learning_rate": 8e-05, "loss": 1.8435, "step": 3594 }, { "epoch": 0.38852264130552255, "grad_norm": 0.3666999638080597, "learning_rate": 8e-05, "loss": 1.6722, "step": 3595 }, { "epoch": 0.3886307143629093, "grad_norm": 0.440245658159256, "learning_rate": 8e-05, "loss": 1.7586, "step": 3596 }, { "epoch": 0.3887387874202961, "grad_norm": 0.4088456928730011, "learning_rate": 8e-05, "loss": 1.6883, "step": 3597 }, { "epoch": 0.3888468604776829, "grad_norm": 0.4411187469959259, "learning_rate": 8e-05, "loss": 1.9444, "step": 3598 }, { "epoch": 0.3889549335350697, "grad_norm": 0.42688512802124023, "learning_rate": 8e-05, "loss": 1.7577, "step": 3599 }, { "epoch": 0.3890630065924565, "grad_norm": 0.4002804160118103, "learning_rate": 8e-05, "loss": 1.6736, "step": 3600 }, { "epoch": 0.3891710796498433, "grad_norm": 0.496926873922348, "learning_rate": 8e-05, "loss": 1.6692, "step": 3601 }, { "epoch": 0.38927915270723007, "grad_norm": 0.369089275598526, "learning_rate": 8e-05, "loss": 1.695, "step": 3602 }, { "epoch": 0.3893872257646169, "grad_norm": 0.3849320113658905, "learning_rate": 8e-05, "loss": 1.7844, "step": 3603 }, { "epoch": 0.3894952988220037, "grad_norm": 0.3766508996486664, "learning_rate": 8e-05, "loss": 1.6605, "step": 3604 }, { "epoch": 0.3896033718793905, "grad_norm": 0.4254394769668579, "learning_rate": 8e-05, "loss": 1.6606, "step": 3605 }, { "epoch": 0.3897114449367773, "grad_norm": 0.3903300166130066, "learning_rate": 8e-05, "loss": 1.7525, "step": 3606 }, { "epoch": 0.38981951799416403, "grad_norm": 0.4021805226802826, "learning_rate": 8e-05, "loss": 1.8197, "step": 3607 }, { "epoch": 0.38992759105155084, "grad_norm": 0.40100282430648804, "learning_rate": 8e-05, "loss": 1.7617, "step": 3608 }, { "epoch": 0.39003566410893764, "grad_norm": 0.398359090089798, "learning_rate": 8e-05, "loss": 1.7468, "step": 3609 }, { "epoch": 0.39014373716632444, "grad_norm": 0.4622156023979187, "learning_rate": 8e-05, "loss": 1.8503, "step": 3610 }, { "epoch": 0.39025181022371125, "grad_norm": 0.40024054050445557, "learning_rate": 8e-05, "loss": 1.7419, "step": 3611 }, { "epoch": 0.390359883281098, "grad_norm": 0.404221773147583, "learning_rate": 8e-05, "loss": 1.4782, "step": 3612 }, { "epoch": 0.3904679563384848, "grad_norm": 0.4110901653766632, "learning_rate": 8e-05, "loss": 1.8794, "step": 3613 }, { "epoch": 0.3905760293958716, "grad_norm": 0.4060975909233093, "learning_rate": 8e-05, "loss": 1.7728, "step": 3614 }, { "epoch": 0.3906841024532584, "grad_norm": 0.3533060848712921, "learning_rate": 8e-05, "loss": 1.6037, "step": 3615 }, { "epoch": 0.3907921755106452, "grad_norm": 0.39177361130714417, "learning_rate": 8e-05, "loss": 1.7524, "step": 3616 }, { "epoch": 0.390900248568032, "grad_norm": 0.3647722899913788, "learning_rate": 8e-05, "loss": 1.812, "step": 3617 }, { "epoch": 0.39100832162541876, "grad_norm": 0.3958182632923126, "learning_rate": 8e-05, "loss": 1.7967, "step": 3618 }, { "epoch": 0.39111639468280557, "grad_norm": 0.383961021900177, "learning_rate": 8e-05, "loss": 1.743, "step": 3619 }, { "epoch": 0.39122446774019237, "grad_norm": 0.4106149971485138, "learning_rate": 8e-05, "loss": 1.7991, "step": 3620 }, { "epoch": 0.3913325407975792, "grad_norm": 0.3814200460910797, "learning_rate": 8e-05, "loss": 1.6136, "step": 3621 }, { "epoch": 0.391440613854966, "grad_norm": 0.42237281799316406, "learning_rate": 8e-05, "loss": 1.7554, "step": 3622 }, { "epoch": 0.39154868691235273, "grad_norm": 0.36670219898223877, "learning_rate": 8e-05, "loss": 1.5479, "step": 3623 }, { "epoch": 0.39165675996973953, "grad_norm": 0.3839063346385956, "learning_rate": 8e-05, "loss": 1.6412, "step": 3624 }, { "epoch": 0.39176483302712634, "grad_norm": 0.3889945149421692, "learning_rate": 8e-05, "loss": 1.6123, "step": 3625 }, { "epoch": 0.39187290608451314, "grad_norm": 0.4028182625770569, "learning_rate": 8e-05, "loss": 1.7791, "step": 3626 }, { "epoch": 0.39198097914189994, "grad_norm": 0.3862074613571167, "learning_rate": 8e-05, "loss": 1.7101, "step": 3627 }, { "epoch": 0.3920890521992867, "grad_norm": 0.3968062102794647, "learning_rate": 8e-05, "loss": 1.6863, "step": 3628 }, { "epoch": 0.3921971252566735, "grad_norm": 0.4119967520236969, "learning_rate": 8e-05, "loss": 1.7788, "step": 3629 }, { "epoch": 0.3923051983140603, "grad_norm": 0.38931092619895935, "learning_rate": 8e-05, "loss": 1.6827, "step": 3630 }, { "epoch": 0.3924132713714471, "grad_norm": 0.39359790086746216, "learning_rate": 8e-05, "loss": 1.7995, "step": 3631 }, { "epoch": 0.3925213444288339, "grad_norm": 0.39946842193603516, "learning_rate": 8e-05, "loss": 1.8105, "step": 3632 }, { "epoch": 0.3926294174862207, "grad_norm": 0.3895476460456848, "learning_rate": 8e-05, "loss": 1.7693, "step": 3633 }, { "epoch": 0.39273749054360746, "grad_norm": 0.4143219292163849, "learning_rate": 8e-05, "loss": 1.7003, "step": 3634 }, { "epoch": 0.39284556360099426, "grad_norm": 0.41944581270217896, "learning_rate": 8e-05, "loss": 1.7369, "step": 3635 }, { "epoch": 0.39295363665838107, "grad_norm": 0.4190118908882141, "learning_rate": 8e-05, "loss": 1.9293, "step": 3636 }, { "epoch": 0.39306170971576787, "grad_norm": 0.40957459807395935, "learning_rate": 8e-05, "loss": 1.845, "step": 3637 }, { "epoch": 0.3931697827731547, "grad_norm": 0.5004032850265503, "learning_rate": 8e-05, "loss": 1.9849, "step": 3638 }, { "epoch": 0.3932778558305414, "grad_norm": 0.38745835423469543, "learning_rate": 8e-05, "loss": 1.8154, "step": 3639 }, { "epoch": 0.3933859288879282, "grad_norm": 0.37838202714920044, "learning_rate": 8e-05, "loss": 1.691, "step": 3640 }, { "epoch": 0.39349400194531503, "grad_norm": 0.4033721387386322, "learning_rate": 8e-05, "loss": 1.8427, "step": 3641 }, { "epoch": 0.39360207500270183, "grad_norm": 0.4197179675102234, "learning_rate": 8e-05, "loss": 1.724, "step": 3642 }, { "epoch": 0.39371014806008864, "grad_norm": 0.42030149698257446, "learning_rate": 8e-05, "loss": 1.6604, "step": 3643 }, { "epoch": 0.3938182211174754, "grad_norm": 0.3908644914627075, "learning_rate": 8e-05, "loss": 1.7046, "step": 3644 }, { "epoch": 0.3939262941748622, "grad_norm": 0.41252681612968445, "learning_rate": 8e-05, "loss": 1.6735, "step": 3645 }, { "epoch": 0.394034367232249, "grad_norm": 0.41587528586387634, "learning_rate": 8e-05, "loss": 1.8748, "step": 3646 }, { "epoch": 0.3941424402896358, "grad_norm": 0.40265074372291565, "learning_rate": 8e-05, "loss": 1.785, "step": 3647 }, { "epoch": 0.3942505133470226, "grad_norm": 0.3616083264350891, "learning_rate": 8e-05, "loss": 1.5792, "step": 3648 }, { "epoch": 0.3943585864044094, "grad_norm": 0.4082064628601074, "learning_rate": 8e-05, "loss": 1.8936, "step": 3649 }, { "epoch": 0.39446665946179615, "grad_norm": 0.38844430446624756, "learning_rate": 8e-05, "loss": 1.7387, "step": 3650 }, { "epoch": 0.39457473251918296, "grad_norm": 0.3799108862876892, "learning_rate": 8e-05, "loss": 1.7097, "step": 3651 }, { "epoch": 0.39468280557656976, "grad_norm": 0.39619314670562744, "learning_rate": 8e-05, "loss": 1.7037, "step": 3652 }, { "epoch": 0.39479087863395657, "grad_norm": 0.4167812764644623, "learning_rate": 8e-05, "loss": 1.8067, "step": 3653 }, { "epoch": 0.39489895169134337, "grad_norm": 0.3990997076034546, "learning_rate": 8e-05, "loss": 1.74, "step": 3654 }, { "epoch": 0.3950070247487301, "grad_norm": 0.4231667220592499, "learning_rate": 8e-05, "loss": 1.7175, "step": 3655 }, { "epoch": 0.3951150978061169, "grad_norm": 0.3799695670604706, "learning_rate": 8e-05, "loss": 1.8591, "step": 3656 }, { "epoch": 0.3952231708635037, "grad_norm": 0.40398335456848145, "learning_rate": 8e-05, "loss": 1.6879, "step": 3657 }, { "epoch": 0.39533124392089053, "grad_norm": 0.43385791778564453, "learning_rate": 8e-05, "loss": 1.7369, "step": 3658 }, { "epoch": 0.39543931697827733, "grad_norm": 0.39882421493530273, "learning_rate": 8e-05, "loss": 1.6642, "step": 3659 }, { "epoch": 0.3955473900356641, "grad_norm": 0.3708348870277405, "learning_rate": 8e-05, "loss": 1.6417, "step": 3660 }, { "epoch": 0.3956554630930509, "grad_norm": 0.36885106563568115, "learning_rate": 8e-05, "loss": 1.4198, "step": 3661 }, { "epoch": 0.3957635361504377, "grad_norm": 0.38646528124809265, "learning_rate": 8e-05, "loss": 1.7347, "step": 3662 }, { "epoch": 0.3958716092078245, "grad_norm": 0.3950575590133667, "learning_rate": 8e-05, "loss": 1.7639, "step": 3663 }, { "epoch": 0.3959796822652113, "grad_norm": 0.40876105427742004, "learning_rate": 8e-05, "loss": 1.6726, "step": 3664 }, { "epoch": 0.3960877553225981, "grad_norm": 0.4051937460899353, "learning_rate": 8e-05, "loss": 1.7237, "step": 3665 }, { "epoch": 0.39619582837998485, "grad_norm": 0.4139687716960907, "learning_rate": 8e-05, "loss": 1.9009, "step": 3666 }, { "epoch": 0.39630390143737165, "grad_norm": 0.3834027647972107, "learning_rate": 8e-05, "loss": 1.7526, "step": 3667 }, { "epoch": 0.39641197449475846, "grad_norm": 0.3897620439529419, "learning_rate": 8e-05, "loss": 1.696, "step": 3668 }, { "epoch": 0.39652004755214526, "grad_norm": 0.41348329186439514, "learning_rate": 8e-05, "loss": 1.6779, "step": 3669 }, { "epoch": 0.39662812060953206, "grad_norm": 0.41085559129714966, "learning_rate": 8e-05, "loss": 1.8176, "step": 3670 }, { "epoch": 0.3967361936669188, "grad_norm": 0.4275372326374054, "learning_rate": 8e-05, "loss": 1.7224, "step": 3671 }, { "epoch": 0.3968442667243056, "grad_norm": 0.36964356899261475, "learning_rate": 8e-05, "loss": 1.6745, "step": 3672 }, { "epoch": 0.3969523397816924, "grad_norm": 0.4059866666793823, "learning_rate": 8e-05, "loss": 1.7421, "step": 3673 }, { "epoch": 0.3970604128390792, "grad_norm": 0.4467698037624359, "learning_rate": 8e-05, "loss": 1.911, "step": 3674 }, { "epoch": 0.39716848589646603, "grad_norm": 0.3793274164199829, "learning_rate": 8e-05, "loss": 1.6528, "step": 3675 }, { "epoch": 0.39727655895385283, "grad_norm": 0.4253424108028412, "learning_rate": 8e-05, "loss": 1.9186, "step": 3676 }, { "epoch": 0.3973846320112396, "grad_norm": 0.3775992691516876, "learning_rate": 8e-05, "loss": 1.4771, "step": 3677 }, { "epoch": 0.3974927050686264, "grad_norm": 0.40357932448387146, "learning_rate": 8e-05, "loss": 1.6427, "step": 3678 }, { "epoch": 0.3976007781260132, "grad_norm": 0.3914742171764374, "learning_rate": 8e-05, "loss": 1.7277, "step": 3679 }, { "epoch": 0.3977088511834, "grad_norm": 0.3934895694255829, "learning_rate": 8e-05, "loss": 1.8489, "step": 3680 }, { "epoch": 0.3978169242407868, "grad_norm": 0.4016740024089813, "learning_rate": 8e-05, "loss": 1.6955, "step": 3681 }, { "epoch": 0.39792499729817354, "grad_norm": 0.40770217776298523, "learning_rate": 8e-05, "loss": 1.7102, "step": 3682 }, { "epoch": 0.39803307035556035, "grad_norm": 0.4512227773666382, "learning_rate": 8e-05, "loss": 1.7325, "step": 3683 }, { "epoch": 0.39814114341294715, "grad_norm": 0.39062368869781494, "learning_rate": 8e-05, "loss": 1.7427, "step": 3684 }, { "epoch": 0.39824921647033396, "grad_norm": 0.49332454800605774, "learning_rate": 8e-05, "loss": 1.843, "step": 3685 }, { "epoch": 0.39835728952772076, "grad_norm": 0.38912585377693176, "learning_rate": 8e-05, "loss": 1.6231, "step": 3686 }, { "epoch": 0.3984653625851075, "grad_norm": 0.37618938088417053, "learning_rate": 8e-05, "loss": 1.6804, "step": 3687 }, { "epoch": 0.3985734356424943, "grad_norm": 0.4008297920227051, "learning_rate": 8e-05, "loss": 1.8939, "step": 3688 }, { "epoch": 0.3986815086998811, "grad_norm": 0.3977849781513214, "learning_rate": 8e-05, "loss": 1.6292, "step": 3689 }, { "epoch": 0.3987895817572679, "grad_norm": 0.38854777812957764, "learning_rate": 8e-05, "loss": 1.6468, "step": 3690 }, { "epoch": 0.3988976548146547, "grad_norm": 0.40625908970832825, "learning_rate": 8e-05, "loss": 1.8091, "step": 3691 }, { "epoch": 0.3990057278720415, "grad_norm": 0.35676607489585876, "learning_rate": 8e-05, "loss": 1.7175, "step": 3692 }, { "epoch": 0.3991138009294283, "grad_norm": 0.3685443103313446, "learning_rate": 8e-05, "loss": 1.6683, "step": 3693 }, { "epoch": 0.3992218739868151, "grad_norm": 0.3897593915462494, "learning_rate": 8e-05, "loss": 1.8239, "step": 3694 }, { "epoch": 0.3993299470442019, "grad_norm": 0.4160090982913971, "learning_rate": 8e-05, "loss": 1.8628, "step": 3695 }, { "epoch": 0.3994380201015887, "grad_norm": 0.4024585485458374, "learning_rate": 8e-05, "loss": 1.5545, "step": 3696 }, { "epoch": 0.3995460931589755, "grad_norm": 0.38592958450317383, "learning_rate": 8e-05, "loss": 1.6371, "step": 3697 }, { "epoch": 0.39965416621636224, "grad_norm": 0.38454344868659973, "learning_rate": 8e-05, "loss": 1.6642, "step": 3698 }, { "epoch": 0.39976223927374904, "grad_norm": 0.431120902299881, "learning_rate": 8e-05, "loss": 1.6842, "step": 3699 }, { "epoch": 0.39987031233113585, "grad_norm": 0.3986557424068451, "learning_rate": 8e-05, "loss": 1.8025, "step": 3700 }, { "epoch": 0.39997838538852265, "grad_norm": 0.4012722373008728, "learning_rate": 8e-05, "loss": 1.8269, "step": 3701 }, { "epoch": 0.40008645844590945, "grad_norm": 0.3755255937576294, "learning_rate": 8e-05, "loss": 1.5911, "step": 3702 }, { "epoch": 0.4001945315032962, "grad_norm": 0.4482969343662262, "learning_rate": 8e-05, "loss": 1.7883, "step": 3703 }, { "epoch": 0.400302604560683, "grad_norm": 0.40132319927215576, "learning_rate": 8e-05, "loss": 1.717, "step": 3704 }, { "epoch": 0.4004106776180698, "grad_norm": 0.376080721616745, "learning_rate": 8e-05, "loss": 1.5695, "step": 3705 }, { "epoch": 0.4005187506754566, "grad_norm": 0.4061757028102875, "learning_rate": 8e-05, "loss": 1.7167, "step": 3706 }, { "epoch": 0.4006268237328434, "grad_norm": 0.41207167506217957, "learning_rate": 8e-05, "loss": 1.7033, "step": 3707 }, { "epoch": 0.4007348967902302, "grad_norm": 0.3914540708065033, "learning_rate": 8e-05, "loss": 1.5986, "step": 3708 }, { "epoch": 0.40084296984761697, "grad_norm": 0.40471914410591125, "learning_rate": 8e-05, "loss": 1.6127, "step": 3709 }, { "epoch": 0.4009510429050038, "grad_norm": 0.4019676446914673, "learning_rate": 8e-05, "loss": 1.6265, "step": 3710 }, { "epoch": 0.4010591159623906, "grad_norm": 0.38933658599853516, "learning_rate": 8e-05, "loss": 1.7484, "step": 3711 }, { "epoch": 0.4011671890197774, "grad_norm": 0.3860172629356384, "learning_rate": 8e-05, "loss": 1.6544, "step": 3712 }, { "epoch": 0.4012752620771642, "grad_norm": 0.3738880455493927, "learning_rate": 8e-05, "loss": 1.6366, "step": 3713 }, { "epoch": 0.40138333513455093, "grad_norm": 0.37177854776382446, "learning_rate": 8e-05, "loss": 1.6966, "step": 3714 }, { "epoch": 0.40149140819193774, "grad_norm": 0.4125951826572418, "learning_rate": 8e-05, "loss": 1.8008, "step": 3715 }, { "epoch": 0.40159948124932454, "grad_norm": 0.40310990810394287, "learning_rate": 8e-05, "loss": 1.677, "step": 3716 }, { "epoch": 0.40170755430671135, "grad_norm": 0.39373481273651123, "learning_rate": 8e-05, "loss": 1.7018, "step": 3717 }, { "epoch": 0.40181562736409815, "grad_norm": 0.40079519152641296, "learning_rate": 8e-05, "loss": 1.7901, "step": 3718 }, { "epoch": 0.4019237004214849, "grad_norm": 0.4064926505088806, "learning_rate": 8e-05, "loss": 1.561, "step": 3719 }, { "epoch": 0.4020317734788717, "grad_norm": 0.4149755835533142, "learning_rate": 8e-05, "loss": 1.7861, "step": 3720 }, { "epoch": 0.4021398465362585, "grad_norm": 0.3788078725337982, "learning_rate": 8e-05, "loss": 1.5809, "step": 3721 }, { "epoch": 0.4022479195936453, "grad_norm": 0.3746579587459564, "learning_rate": 8e-05, "loss": 1.5581, "step": 3722 }, { "epoch": 0.4023559926510321, "grad_norm": 0.39097830653190613, "learning_rate": 8e-05, "loss": 1.7712, "step": 3723 }, { "epoch": 0.4024640657084189, "grad_norm": 0.3641214966773987, "learning_rate": 8e-05, "loss": 1.6824, "step": 3724 }, { "epoch": 0.40257213876580566, "grad_norm": 0.3794381320476532, "learning_rate": 8e-05, "loss": 1.8231, "step": 3725 }, { "epoch": 0.40268021182319247, "grad_norm": 0.3842198848724365, "learning_rate": 8e-05, "loss": 1.7266, "step": 3726 }, { "epoch": 0.4027882848805793, "grad_norm": 0.39962583780288696, "learning_rate": 8e-05, "loss": 1.8617, "step": 3727 }, { "epoch": 0.4028963579379661, "grad_norm": 0.40599000453948975, "learning_rate": 8e-05, "loss": 1.6513, "step": 3728 }, { "epoch": 0.4030044309953529, "grad_norm": 0.4413057863712311, "learning_rate": 8e-05, "loss": 1.8482, "step": 3729 }, { "epoch": 0.40311250405273963, "grad_norm": 0.377800315618515, "learning_rate": 8e-05, "loss": 1.6702, "step": 3730 }, { "epoch": 0.40322057711012643, "grad_norm": 0.39248013496398926, "learning_rate": 8e-05, "loss": 1.6857, "step": 3731 }, { "epoch": 0.40332865016751324, "grad_norm": 0.38526687026023865, "learning_rate": 8e-05, "loss": 1.5928, "step": 3732 }, { "epoch": 0.40343672322490004, "grad_norm": 0.44204992055892944, "learning_rate": 8e-05, "loss": 1.9811, "step": 3733 }, { "epoch": 0.40354479628228684, "grad_norm": 0.43923699855804443, "learning_rate": 8e-05, "loss": 1.8502, "step": 3734 }, { "epoch": 0.4036528693396736, "grad_norm": 0.3984847068786621, "learning_rate": 8e-05, "loss": 1.6276, "step": 3735 }, { "epoch": 0.4037609423970604, "grad_norm": 0.4320618212223053, "learning_rate": 8e-05, "loss": 1.9047, "step": 3736 }, { "epoch": 0.4038690154544472, "grad_norm": 0.36007046699523926, "learning_rate": 8e-05, "loss": 1.5466, "step": 3737 }, { "epoch": 0.403977088511834, "grad_norm": 0.38590195775032043, "learning_rate": 8e-05, "loss": 1.8055, "step": 3738 }, { "epoch": 0.4040851615692208, "grad_norm": 0.39130690693855286, "learning_rate": 8e-05, "loss": 1.6736, "step": 3739 }, { "epoch": 0.4041932346266076, "grad_norm": 0.4166002869606018, "learning_rate": 8e-05, "loss": 1.9243, "step": 3740 }, { "epoch": 0.40430130768399436, "grad_norm": 0.376675009727478, "learning_rate": 8e-05, "loss": 1.5765, "step": 3741 }, { "epoch": 0.40440938074138116, "grad_norm": 0.3769833743572235, "learning_rate": 8e-05, "loss": 1.6408, "step": 3742 }, { "epoch": 0.40451745379876797, "grad_norm": 0.384000301361084, "learning_rate": 8e-05, "loss": 1.7887, "step": 3743 }, { "epoch": 0.40462552685615477, "grad_norm": 0.3945089280605316, "learning_rate": 8e-05, "loss": 1.7568, "step": 3744 }, { "epoch": 0.4047335999135416, "grad_norm": 0.393502801656723, "learning_rate": 8e-05, "loss": 1.8339, "step": 3745 }, { "epoch": 0.4048416729709283, "grad_norm": 0.39680758118629456, "learning_rate": 8e-05, "loss": 1.7304, "step": 3746 }, { "epoch": 0.4049497460283151, "grad_norm": 0.38890594244003296, "learning_rate": 8e-05, "loss": 1.5778, "step": 3747 }, { "epoch": 0.40505781908570193, "grad_norm": 0.35278794169425964, "learning_rate": 8e-05, "loss": 1.6004, "step": 3748 }, { "epoch": 0.40516589214308874, "grad_norm": 0.3913782835006714, "learning_rate": 8e-05, "loss": 1.63, "step": 3749 }, { "epoch": 0.40527396520047554, "grad_norm": 0.37278279662132263, "learning_rate": 8e-05, "loss": 1.7734, "step": 3750 }, { "epoch": 0.4053820382578623, "grad_norm": 0.39850932359695435, "learning_rate": 8e-05, "loss": 1.4858, "step": 3751 }, { "epoch": 0.4054901113152491, "grad_norm": 0.3929099440574646, "learning_rate": 8e-05, "loss": 1.7001, "step": 3752 }, { "epoch": 0.4055981843726359, "grad_norm": 0.42637261748313904, "learning_rate": 8e-05, "loss": 1.7951, "step": 3753 }, { "epoch": 0.4057062574300227, "grad_norm": 0.3746306300163269, "learning_rate": 8e-05, "loss": 1.626, "step": 3754 }, { "epoch": 0.4058143304874095, "grad_norm": 0.38104248046875, "learning_rate": 8e-05, "loss": 1.6227, "step": 3755 }, { "epoch": 0.4059224035447963, "grad_norm": 0.37350109219551086, "learning_rate": 8e-05, "loss": 1.6684, "step": 3756 }, { "epoch": 0.40603047660218305, "grad_norm": 0.39378634095191956, "learning_rate": 8e-05, "loss": 1.7541, "step": 3757 }, { "epoch": 0.40613854965956986, "grad_norm": 0.3619995415210724, "learning_rate": 8e-05, "loss": 1.5679, "step": 3758 }, { "epoch": 0.40624662271695666, "grad_norm": 0.41921520233154297, "learning_rate": 8e-05, "loss": 1.7098, "step": 3759 }, { "epoch": 0.40635469577434347, "grad_norm": 0.3966827392578125, "learning_rate": 8e-05, "loss": 1.757, "step": 3760 }, { "epoch": 0.40646276883173027, "grad_norm": 0.41021886467933655, "learning_rate": 8e-05, "loss": 1.7714, "step": 3761 }, { "epoch": 0.406570841889117, "grad_norm": 0.40492698550224304, "learning_rate": 8e-05, "loss": 1.7581, "step": 3762 }, { "epoch": 0.4066789149465038, "grad_norm": 0.427545964717865, "learning_rate": 8e-05, "loss": 1.6431, "step": 3763 }, { "epoch": 0.4067869880038906, "grad_norm": 0.42807745933532715, "learning_rate": 8e-05, "loss": 1.7861, "step": 3764 }, { "epoch": 0.40689506106127743, "grad_norm": 0.39299434423446655, "learning_rate": 8e-05, "loss": 1.8002, "step": 3765 }, { "epoch": 0.40700313411866423, "grad_norm": 0.3910335600376129, "learning_rate": 8e-05, "loss": 1.5639, "step": 3766 }, { "epoch": 0.40711120717605104, "grad_norm": 0.4081684350967407, "learning_rate": 8e-05, "loss": 1.6153, "step": 3767 }, { "epoch": 0.4072192802334378, "grad_norm": 0.3840066194534302, "learning_rate": 8e-05, "loss": 1.6344, "step": 3768 }, { "epoch": 0.4073273532908246, "grad_norm": 0.3912813067436218, "learning_rate": 8e-05, "loss": 1.6825, "step": 3769 }, { "epoch": 0.4074354263482114, "grad_norm": 0.3840087652206421, "learning_rate": 8e-05, "loss": 1.7064, "step": 3770 }, { "epoch": 0.4075434994055982, "grad_norm": 0.38698792457580566, "learning_rate": 8e-05, "loss": 1.7397, "step": 3771 }, { "epoch": 0.407651572462985, "grad_norm": 0.3900686204433441, "learning_rate": 8e-05, "loss": 1.555, "step": 3772 }, { "epoch": 0.40775964552037175, "grad_norm": 0.3696495592594147, "learning_rate": 8e-05, "loss": 1.6111, "step": 3773 }, { "epoch": 0.40786771857775855, "grad_norm": 0.3842237889766693, "learning_rate": 8e-05, "loss": 1.739, "step": 3774 }, { "epoch": 0.40797579163514536, "grad_norm": 0.38225430250167847, "learning_rate": 8e-05, "loss": 1.6885, "step": 3775 }, { "epoch": 0.40808386469253216, "grad_norm": 0.3904886245727539, "learning_rate": 8e-05, "loss": 1.8409, "step": 3776 }, { "epoch": 0.40819193774991896, "grad_norm": 0.3797644376754761, "learning_rate": 8e-05, "loss": 1.7335, "step": 3777 }, { "epoch": 0.4083000108073057, "grad_norm": 0.39331331849098206, "learning_rate": 8e-05, "loss": 1.6978, "step": 3778 }, { "epoch": 0.4084080838646925, "grad_norm": 0.41202792525291443, "learning_rate": 8e-05, "loss": 1.8553, "step": 3779 }, { "epoch": 0.4085161569220793, "grad_norm": 0.39176079630851746, "learning_rate": 8e-05, "loss": 1.7069, "step": 3780 }, { "epoch": 0.4086242299794661, "grad_norm": 0.41179949045181274, "learning_rate": 8e-05, "loss": 1.8179, "step": 3781 }, { "epoch": 0.40873230303685293, "grad_norm": 0.3702789545059204, "learning_rate": 8e-05, "loss": 1.5119, "step": 3782 }, { "epoch": 0.40884037609423973, "grad_norm": 0.38579022884368896, "learning_rate": 8e-05, "loss": 1.7276, "step": 3783 }, { "epoch": 0.4089484491516265, "grad_norm": 0.44669055938720703, "learning_rate": 8e-05, "loss": 1.9425, "step": 3784 }, { "epoch": 0.4090565222090133, "grad_norm": 0.41179758310317993, "learning_rate": 8e-05, "loss": 1.7262, "step": 3785 }, { "epoch": 0.4091645952664001, "grad_norm": 0.4036073684692383, "learning_rate": 8e-05, "loss": 1.7531, "step": 3786 }, { "epoch": 0.4092726683237869, "grad_norm": 0.41674765944480896, "learning_rate": 8e-05, "loss": 1.8272, "step": 3787 }, { "epoch": 0.4093807413811737, "grad_norm": 0.42446672916412354, "learning_rate": 8e-05, "loss": 1.8081, "step": 3788 }, { "epoch": 0.40948881443856044, "grad_norm": 0.395670086145401, "learning_rate": 8e-05, "loss": 1.72, "step": 3789 }, { "epoch": 0.40959688749594725, "grad_norm": 0.40348920226097107, "learning_rate": 8e-05, "loss": 1.7283, "step": 3790 }, { "epoch": 0.40970496055333405, "grad_norm": 0.3769279420375824, "learning_rate": 8e-05, "loss": 1.7197, "step": 3791 }, { "epoch": 0.40981303361072086, "grad_norm": 0.381083607673645, "learning_rate": 8e-05, "loss": 1.8504, "step": 3792 }, { "epoch": 0.40992110666810766, "grad_norm": 0.39807000756263733, "learning_rate": 8e-05, "loss": 1.7388, "step": 3793 }, { "epoch": 0.4100291797254944, "grad_norm": 0.4081137776374817, "learning_rate": 8e-05, "loss": 1.7205, "step": 3794 }, { "epoch": 0.4101372527828812, "grad_norm": 0.43015384674072266, "learning_rate": 8e-05, "loss": 1.7789, "step": 3795 }, { "epoch": 0.410245325840268, "grad_norm": 0.4252828061580658, "learning_rate": 8e-05, "loss": 1.7749, "step": 3796 }, { "epoch": 0.4103533988976548, "grad_norm": 0.4680314064025879, "learning_rate": 8e-05, "loss": 1.7288, "step": 3797 }, { "epoch": 0.4104614719550416, "grad_norm": 0.4137987494468689, "learning_rate": 8e-05, "loss": 1.7295, "step": 3798 }, { "epoch": 0.4105695450124284, "grad_norm": 0.40972837805747986, "learning_rate": 8e-05, "loss": 1.9828, "step": 3799 }, { "epoch": 0.4106776180698152, "grad_norm": 0.4470110833644867, "learning_rate": 8e-05, "loss": 1.8219, "step": 3800 }, { "epoch": 0.410785691127202, "grad_norm": 0.40168553590774536, "learning_rate": 8e-05, "loss": 1.7093, "step": 3801 }, { "epoch": 0.4108937641845888, "grad_norm": 0.3665669560432434, "learning_rate": 8e-05, "loss": 1.6479, "step": 3802 }, { "epoch": 0.4110018372419756, "grad_norm": 0.37971702218055725, "learning_rate": 8e-05, "loss": 1.5925, "step": 3803 }, { "epoch": 0.4111099102993624, "grad_norm": 0.36938971281051636, "learning_rate": 8e-05, "loss": 1.6143, "step": 3804 }, { "epoch": 0.41121798335674914, "grad_norm": 0.40479499101638794, "learning_rate": 8e-05, "loss": 1.7483, "step": 3805 }, { "epoch": 0.41132605641413594, "grad_norm": 0.3785454332828522, "learning_rate": 8e-05, "loss": 1.6935, "step": 3806 }, { "epoch": 0.41143412947152275, "grad_norm": 0.42723578214645386, "learning_rate": 8e-05, "loss": 1.8046, "step": 3807 }, { "epoch": 0.41154220252890955, "grad_norm": 0.37585800886154175, "learning_rate": 8e-05, "loss": 1.645, "step": 3808 }, { "epoch": 0.41165027558629635, "grad_norm": 0.43754732608795166, "learning_rate": 8e-05, "loss": 1.9067, "step": 3809 }, { "epoch": 0.4117583486436831, "grad_norm": 0.36948803067207336, "learning_rate": 8e-05, "loss": 1.6548, "step": 3810 }, { "epoch": 0.4118664217010699, "grad_norm": 0.37926816940307617, "learning_rate": 8e-05, "loss": 1.6941, "step": 3811 }, { "epoch": 0.4119744947584567, "grad_norm": 0.3859090805053711, "learning_rate": 8e-05, "loss": 1.761, "step": 3812 }, { "epoch": 0.4120825678158435, "grad_norm": 0.4168667793273926, "learning_rate": 8e-05, "loss": 1.6544, "step": 3813 }, { "epoch": 0.4121906408732303, "grad_norm": 0.3852517008781433, "learning_rate": 8e-05, "loss": 1.8059, "step": 3814 }, { "epoch": 0.4122987139306171, "grad_norm": 0.3861292004585266, "learning_rate": 8e-05, "loss": 1.692, "step": 3815 }, { "epoch": 0.41240678698800387, "grad_norm": 0.40395817160606384, "learning_rate": 8e-05, "loss": 1.6503, "step": 3816 }, { "epoch": 0.4125148600453907, "grad_norm": 0.38121846318244934, "learning_rate": 8e-05, "loss": 1.5806, "step": 3817 }, { "epoch": 0.4126229331027775, "grad_norm": 0.4147404134273529, "learning_rate": 8e-05, "loss": 1.7027, "step": 3818 }, { "epoch": 0.4127310061601643, "grad_norm": 0.3901894986629486, "learning_rate": 8e-05, "loss": 1.5535, "step": 3819 }, { "epoch": 0.4128390792175511, "grad_norm": 0.4096665382385254, "learning_rate": 8e-05, "loss": 1.8196, "step": 3820 }, { "epoch": 0.41294715227493783, "grad_norm": 0.3848898708820343, "learning_rate": 8e-05, "loss": 1.5894, "step": 3821 }, { "epoch": 0.41305522533232464, "grad_norm": 0.4184325635433197, "learning_rate": 8e-05, "loss": 1.6914, "step": 3822 }, { "epoch": 0.41316329838971144, "grad_norm": 0.4055677354335785, "learning_rate": 8e-05, "loss": 1.6885, "step": 3823 }, { "epoch": 0.41327137144709825, "grad_norm": 0.42235785722732544, "learning_rate": 8e-05, "loss": 1.7978, "step": 3824 }, { "epoch": 0.41337944450448505, "grad_norm": 0.41004014015197754, "learning_rate": 8e-05, "loss": 1.8054, "step": 3825 }, { "epoch": 0.4134875175618718, "grad_norm": 0.4179190397262573, "learning_rate": 8e-05, "loss": 1.7072, "step": 3826 }, { "epoch": 0.4135955906192586, "grad_norm": 0.3941066265106201, "learning_rate": 8e-05, "loss": 1.4534, "step": 3827 }, { "epoch": 0.4137036636766454, "grad_norm": 0.4446079730987549, "learning_rate": 8e-05, "loss": 1.5991, "step": 3828 }, { "epoch": 0.4138117367340322, "grad_norm": 0.39789435267448425, "learning_rate": 8e-05, "loss": 1.3593, "step": 3829 }, { "epoch": 0.413919809791419, "grad_norm": 0.4304432272911072, "learning_rate": 8e-05, "loss": 1.9247, "step": 3830 }, { "epoch": 0.4140278828488058, "grad_norm": 0.38246092200279236, "learning_rate": 8e-05, "loss": 1.6697, "step": 3831 }, { "epoch": 0.41413595590619257, "grad_norm": 0.4152018129825592, "learning_rate": 8e-05, "loss": 1.7623, "step": 3832 }, { "epoch": 0.41424402896357937, "grad_norm": 0.3686734139919281, "learning_rate": 8e-05, "loss": 1.5994, "step": 3833 }, { "epoch": 0.4143521020209662, "grad_norm": 0.3893820345401764, "learning_rate": 8e-05, "loss": 1.7034, "step": 3834 }, { "epoch": 0.414460175078353, "grad_norm": 0.4289054572582245, "learning_rate": 8e-05, "loss": 1.733, "step": 3835 }, { "epoch": 0.4145682481357398, "grad_norm": 0.4365761876106262, "learning_rate": 8e-05, "loss": 1.7199, "step": 3836 }, { "epoch": 0.41467632119312653, "grad_norm": 0.40079325437545776, "learning_rate": 8e-05, "loss": 1.7511, "step": 3837 }, { "epoch": 0.41478439425051333, "grad_norm": 0.43482568860054016, "learning_rate": 8e-05, "loss": 1.8854, "step": 3838 }, { "epoch": 0.41489246730790014, "grad_norm": 0.3957137167453766, "learning_rate": 8e-05, "loss": 1.7392, "step": 3839 }, { "epoch": 0.41500054036528694, "grad_norm": 0.3891526460647583, "learning_rate": 8e-05, "loss": 1.8902, "step": 3840 }, { "epoch": 0.41510861342267374, "grad_norm": 0.36817070841789246, "learning_rate": 8e-05, "loss": 1.6249, "step": 3841 }, { "epoch": 0.41521668648006055, "grad_norm": 0.4674880802631378, "learning_rate": 8e-05, "loss": 2.1006, "step": 3842 }, { "epoch": 0.4153247595374473, "grad_norm": 0.3756575286388397, "learning_rate": 8e-05, "loss": 1.5585, "step": 3843 }, { "epoch": 0.4154328325948341, "grad_norm": 0.42478275299072266, "learning_rate": 8e-05, "loss": 1.8408, "step": 3844 }, { "epoch": 0.4155409056522209, "grad_norm": 0.42642882466316223, "learning_rate": 8e-05, "loss": 1.7534, "step": 3845 }, { "epoch": 0.4156489787096077, "grad_norm": 0.4265595078468323, "learning_rate": 8e-05, "loss": 1.7018, "step": 3846 }, { "epoch": 0.4157570517669945, "grad_norm": 0.4081973135471344, "learning_rate": 8e-05, "loss": 1.7751, "step": 3847 }, { "epoch": 0.41586512482438126, "grad_norm": 0.38063693046569824, "learning_rate": 8e-05, "loss": 1.7751, "step": 3848 }, { "epoch": 0.41597319788176806, "grad_norm": 0.4516310691833496, "learning_rate": 8e-05, "loss": 2.0244, "step": 3849 }, { "epoch": 0.41608127093915487, "grad_norm": 0.4231090843677521, "learning_rate": 8e-05, "loss": 1.7643, "step": 3850 }, { "epoch": 0.41618934399654167, "grad_norm": 0.36689531803131104, "learning_rate": 8e-05, "loss": 1.616, "step": 3851 }, { "epoch": 0.4162974170539285, "grad_norm": 0.42786717414855957, "learning_rate": 8e-05, "loss": 1.8138, "step": 3852 }, { "epoch": 0.4164054901113152, "grad_norm": 0.34821581840515137, "learning_rate": 8e-05, "loss": 1.563, "step": 3853 }, { "epoch": 0.41651356316870203, "grad_norm": 0.38966771960258484, "learning_rate": 8e-05, "loss": 1.8323, "step": 3854 }, { "epoch": 0.41662163622608883, "grad_norm": 0.43030041456222534, "learning_rate": 8e-05, "loss": 1.6704, "step": 3855 }, { "epoch": 0.41672970928347564, "grad_norm": 0.4543220102787018, "learning_rate": 8e-05, "loss": 1.9392, "step": 3856 }, { "epoch": 0.41683778234086244, "grad_norm": 0.3875998556613922, "learning_rate": 8e-05, "loss": 1.5342, "step": 3857 }, { "epoch": 0.41694585539824924, "grad_norm": 0.40602919459342957, "learning_rate": 8e-05, "loss": 1.748, "step": 3858 }, { "epoch": 0.417053928455636, "grad_norm": 0.4219368100166321, "learning_rate": 8e-05, "loss": 1.8748, "step": 3859 }, { "epoch": 0.4171620015130228, "grad_norm": 0.41793370246887207, "learning_rate": 8e-05, "loss": 1.7466, "step": 3860 }, { "epoch": 0.4172700745704096, "grad_norm": 0.4360904395580292, "learning_rate": 8e-05, "loss": 1.9035, "step": 3861 }, { "epoch": 0.4173781476277964, "grad_norm": 0.41942670941352844, "learning_rate": 8e-05, "loss": 1.9133, "step": 3862 }, { "epoch": 0.4174862206851832, "grad_norm": 0.40040525794029236, "learning_rate": 8e-05, "loss": 1.6959, "step": 3863 }, { "epoch": 0.41759429374256996, "grad_norm": 0.3819044530391693, "learning_rate": 8e-05, "loss": 1.7643, "step": 3864 }, { "epoch": 0.41770236679995676, "grad_norm": 0.42044445872306824, "learning_rate": 8e-05, "loss": 1.8157, "step": 3865 }, { "epoch": 0.41781043985734356, "grad_norm": 0.36946722865104675, "learning_rate": 8e-05, "loss": 1.6438, "step": 3866 }, { "epoch": 0.41791851291473037, "grad_norm": 0.4200522005558014, "learning_rate": 8e-05, "loss": 1.73, "step": 3867 }, { "epoch": 0.41802658597211717, "grad_norm": 0.3917948603630066, "learning_rate": 8e-05, "loss": 1.7759, "step": 3868 }, { "epoch": 0.4181346590295039, "grad_norm": 0.4072740077972412, "learning_rate": 8e-05, "loss": 1.7675, "step": 3869 }, { "epoch": 0.4182427320868907, "grad_norm": 0.3877619504928589, "learning_rate": 8e-05, "loss": 1.6873, "step": 3870 }, { "epoch": 0.4183508051442775, "grad_norm": 0.4184662699699402, "learning_rate": 8e-05, "loss": 1.8491, "step": 3871 }, { "epoch": 0.41845887820166433, "grad_norm": 0.3985510468482971, "learning_rate": 8e-05, "loss": 1.7071, "step": 3872 }, { "epoch": 0.41856695125905113, "grad_norm": 0.39273327589035034, "learning_rate": 8e-05, "loss": 1.7745, "step": 3873 }, { "epoch": 0.41867502431643794, "grad_norm": 0.3655044138431549, "learning_rate": 8e-05, "loss": 1.5411, "step": 3874 }, { "epoch": 0.4187830973738247, "grad_norm": 0.3853136897087097, "learning_rate": 8e-05, "loss": 1.7062, "step": 3875 }, { "epoch": 0.4188911704312115, "grad_norm": 0.43215006589889526, "learning_rate": 8e-05, "loss": 1.8042, "step": 3876 }, { "epoch": 0.4189992434885983, "grad_norm": 0.382182776927948, "learning_rate": 8e-05, "loss": 1.7719, "step": 3877 }, { "epoch": 0.4191073165459851, "grad_norm": 0.4090704619884491, "learning_rate": 8e-05, "loss": 1.4606, "step": 3878 }, { "epoch": 0.4192153896033719, "grad_norm": 0.37792569398880005, "learning_rate": 8e-05, "loss": 1.8337, "step": 3879 }, { "epoch": 0.41932346266075865, "grad_norm": 0.3886340856552124, "learning_rate": 8e-05, "loss": 1.7306, "step": 3880 }, { "epoch": 0.41943153571814545, "grad_norm": 0.39856937527656555, "learning_rate": 8e-05, "loss": 1.7665, "step": 3881 }, { "epoch": 0.41953960877553226, "grad_norm": 0.42731624841690063, "learning_rate": 8e-05, "loss": 1.7501, "step": 3882 }, { "epoch": 0.41964768183291906, "grad_norm": 0.4575856328010559, "learning_rate": 8e-05, "loss": 1.5855, "step": 3883 }, { "epoch": 0.41975575489030587, "grad_norm": 0.3995247781276703, "learning_rate": 8e-05, "loss": 1.793, "step": 3884 }, { "epoch": 0.4198638279476926, "grad_norm": 0.4094035029411316, "learning_rate": 8e-05, "loss": 1.8229, "step": 3885 }, { "epoch": 0.4199719010050794, "grad_norm": 0.3868167996406555, "learning_rate": 8e-05, "loss": 1.4925, "step": 3886 }, { "epoch": 0.4200799740624662, "grad_norm": 0.3684050738811493, "learning_rate": 8e-05, "loss": 1.6274, "step": 3887 }, { "epoch": 0.420188047119853, "grad_norm": 0.48412683606147766, "learning_rate": 8e-05, "loss": 1.9776, "step": 3888 }, { "epoch": 0.42029612017723983, "grad_norm": 0.3780952990055084, "learning_rate": 8e-05, "loss": 1.6968, "step": 3889 }, { "epoch": 0.42040419323462663, "grad_norm": 0.363110214471817, "learning_rate": 8e-05, "loss": 1.7397, "step": 3890 }, { "epoch": 0.4205122662920134, "grad_norm": 0.43607765436172485, "learning_rate": 8e-05, "loss": 1.7425, "step": 3891 }, { "epoch": 0.4206203393494002, "grad_norm": 0.3985387980937958, "learning_rate": 8e-05, "loss": 1.8339, "step": 3892 }, { "epoch": 0.420728412406787, "grad_norm": 0.40517929196357727, "learning_rate": 8e-05, "loss": 1.9024, "step": 3893 }, { "epoch": 0.4208364854641738, "grad_norm": 0.3849782943725586, "learning_rate": 8e-05, "loss": 1.6044, "step": 3894 }, { "epoch": 0.4209445585215606, "grad_norm": 0.38462749123573303, "learning_rate": 8e-05, "loss": 1.7033, "step": 3895 }, { "epoch": 0.42105263157894735, "grad_norm": 0.38256916403770447, "learning_rate": 8e-05, "loss": 1.5715, "step": 3896 }, { "epoch": 0.42116070463633415, "grad_norm": 0.3805573284626007, "learning_rate": 8e-05, "loss": 1.6983, "step": 3897 }, { "epoch": 0.42126877769372095, "grad_norm": 0.4165429472923279, "learning_rate": 8e-05, "loss": 1.7593, "step": 3898 }, { "epoch": 0.42137685075110776, "grad_norm": 0.4750782549381256, "learning_rate": 8e-05, "loss": 1.8399, "step": 3899 }, { "epoch": 0.42148492380849456, "grad_norm": 0.4078114330768585, "learning_rate": 8e-05, "loss": 1.7572, "step": 3900 }, { "epoch": 0.4215929968658813, "grad_norm": 0.3825790584087372, "learning_rate": 8e-05, "loss": 1.547, "step": 3901 }, { "epoch": 0.4217010699232681, "grad_norm": 0.42685189843177795, "learning_rate": 8e-05, "loss": 1.7911, "step": 3902 }, { "epoch": 0.4218091429806549, "grad_norm": 0.3979296088218689, "learning_rate": 8e-05, "loss": 1.6465, "step": 3903 }, { "epoch": 0.4219172160380417, "grad_norm": 0.38864952325820923, "learning_rate": 8e-05, "loss": 1.7216, "step": 3904 }, { "epoch": 0.4220252890954285, "grad_norm": 0.4438540041446686, "learning_rate": 8e-05, "loss": 1.8995, "step": 3905 }, { "epoch": 0.42213336215281533, "grad_norm": 0.41845783591270447, "learning_rate": 8e-05, "loss": 1.7492, "step": 3906 }, { "epoch": 0.4222414352102021, "grad_norm": 0.41742652654647827, "learning_rate": 8e-05, "loss": 1.6932, "step": 3907 }, { "epoch": 0.4223495082675889, "grad_norm": 0.4258211851119995, "learning_rate": 8e-05, "loss": 1.9198, "step": 3908 }, { "epoch": 0.4224575813249757, "grad_norm": 0.37889838218688965, "learning_rate": 8e-05, "loss": 1.4833, "step": 3909 }, { "epoch": 0.4225656543823625, "grad_norm": 0.39772534370422363, "learning_rate": 8e-05, "loss": 1.7301, "step": 3910 }, { "epoch": 0.4226737274397493, "grad_norm": 0.37279587984085083, "learning_rate": 8e-05, "loss": 1.7302, "step": 3911 }, { "epoch": 0.42278180049713604, "grad_norm": 0.4219800531864166, "learning_rate": 8e-05, "loss": 1.8416, "step": 3912 }, { "epoch": 0.42288987355452284, "grad_norm": 0.4103454053401947, "learning_rate": 8e-05, "loss": 1.7689, "step": 3913 }, { "epoch": 0.42299794661190965, "grad_norm": 0.4265066385269165, "learning_rate": 8e-05, "loss": 1.8538, "step": 3914 }, { "epoch": 0.42310601966929645, "grad_norm": 0.37320035696029663, "learning_rate": 8e-05, "loss": 1.6804, "step": 3915 }, { "epoch": 0.42321409272668326, "grad_norm": 0.3801858425140381, "learning_rate": 8e-05, "loss": 1.6682, "step": 3916 }, { "epoch": 0.42332216578407, "grad_norm": 0.4436030089855194, "learning_rate": 8e-05, "loss": 1.9305, "step": 3917 }, { "epoch": 0.4234302388414568, "grad_norm": 0.37914249300956726, "learning_rate": 8e-05, "loss": 1.6613, "step": 3918 }, { "epoch": 0.4235383118988436, "grad_norm": 0.3867349922657013, "learning_rate": 8e-05, "loss": 1.6651, "step": 3919 }, { "epoch": 0.4236463849562304, "grad_norm": 0.416350394487381, "learning_rate": 8e-05, "loss": 1.8336, "step": 3920 }, { "epoch": 0.4237544580136172, "grad_norm": 0.39160364866256714, "learning_rate": 8e-05, "loss": 1.6501, "step": 3921 }, { "epoch": 0.423862531071004, "grad_norm": 0.44497206807136536, "learning_rate": 8e-05, "loss": 1.789, "step": 3922 }, { "epoch": 0.42397060412839077, "grad_norm": 0.395686537027359, "learning_rate": 8e-05, "loss": 1.6371, "step": 3923 }, { "epoch": 0.4240786771857776, "grad_norm": 0.3818310499191284, "learning_rate": 8e-05, "loss": 1.591, "step": 3924 }, { "epoch": 0.4241867502431644, "grad_norm": 0.3982886075973511, "learning_rate": 8e-05, "loss": 1.8616, "step": 3925 }, { "epoch": 0.4242948233005512, "grad_norm": 0.3833780884742737, "learning_rate": 8e-05, "loss": 1.737, "step": 3926 }, { "epoch": 0.424402896357938, "grad_norm": 0.39365801215171814, "learning_rate": 8e-05, "loss": 1.777, "step": 3927 }, { "epoch": 0.42451096941532473, "grad_norm": 0.4016670286655426, "learning_rate": 8e-05, "loss": 1.7574, "step": 3928 }, { "epoch": 0.42461904247271154, "grad_norm": 0.3761076033115387, "learning_rate": 8e-05, "loss": 1.732, "step": 3929 }, { "epoch": 0.42472711553009834, "grad_norm": 0.3949795365333557, "learning_rate": 8e-05, "loss": 1.659, "step": 3930 }, { "epoch": 0.42483518858748515, "grad_norm": 0.4117641746997833, "learning_rate": 8e-05, "loss": 1.8549, "step": 3931 }, { "epoch": 0.42494326164487195, "grad_norm": 0.44422808289527893, "learning_rate": 8e-05, "loss": 1.8429, "step": 3932 }, { "epoch": 0.42505133470225875, "grad_norm": 0.38675034046173096, "learning_rate": 8e-05, "loss": 1.5983, "step": 3933 }, { "epoch": 0.4251594077596455, "grad_norm": 0.46386539936065674, "learning_rate": 8e-05, "loss": 1.7914, "step": 3934 }, { "epoch": 0.4252674808170323, "grad_norm": 0.38414624333381653, "learning_rate": 8e-05, "loss": 1.6158, "step": 3935 }, { "epoch": 0.4253755538744191, "grad_norm": 0.42299190163612366, "learning_rate": 8e-05, "loss": 1.7692, "step": 3936 }, { "epoch": 0.4254836269318059, "grad_norm": 0.39568471908569336, "learning_rate": 8e-05, "loss": 1.6804, "step": 3937 }, { "epoch": 0.4255916999891927, "grad_norm": 0.4121760129928589, "learning_rate": 8e-05, "loss": 1.7699, "step": 3938 }, { "epoch": 0.42569977304657947, "grad_norm": 0.4374620020389557, "learning_rate": 8e-05, "loss": 1.5819, "step": 3939 }, { "epoch": 0.42580784610396627, "grad_norm": 0.4322124421596527, "learning_rate": 8e-05, "loss": 1.8521, "step": 3940 }, { "epoch": 0.4259159191613531, "grad_norm": 0.38904377818107605, "learning_rate": 8e-05, "loss": 1.7714, "step": 3941 }, { "epoch": 0.4260239922187399, "grad_norm": 0.35896697640419006, "learning_rate": 8e-05, "loss": 1.576, "step": 3942 }, { "epoch": 0.4261320652761267, "grad_norm": 0.41874027252197266, "learning_rate": 8e-05, "loss": 1.7153, "step": 3943 }, { "epoch": 0.42624013833351343, "grad_norm": 0.3961946964263916, "learning_rate": 8e-05, "loss": 1.794, "step": 3944 }, { "epoch": 0.42634821139090023, "grad_norm": 0.39152586460113525, "learning_rate": 8e-05, "loss": 1.67, "step": 3945 }, { "epoch": 0.42645628444828704, "grad_norm": 0.36863604187965393, "learning_rate": 8e-05, "loss": 1.6239, "step": 3946 }, { "epoch": 0.42656435750567384, "grad_norm": 0.4661109447479248, "learning_rate": 8e-05, "loss": 1.9955, "step": 3947 }, { "epoch": 0.42667243056306065, "grad_norm": 0.44990408420562744, "learning_rate": 8e-05, "loss": 1.9217, "step": 3948 }, { "epoch": 0.42678050362044745, "grad_norm": 0.39468148350715637, "learning_rate": 8e-05, "loss": 1.7014, "step": 3949 }, { "epoch": 0.4268885766778342, "grad_norm": 0.375744491815567, "learning_rate": 8e-05, "loss": 1.6788, "step": 3950 }, { "epoch": 0.426996649735221, "grad_norm": 0.37918198108673096, "learning_rate": 8e-05, "loss": 1.6259, "step": 3951 }, { "epoch": 0.4271047227926078, "grad_norm": 0.41411536931991577, "learning_rate": 8e-05, "loss": 1.7507, "step": 3952 }, { "epoch": 0.4272127958499946, "grad_norm": 0.4107980728149414, "learning_rate": 8e-05, "loss": 1.9219, "step": 3953 }, { "epoch": 0.4273208689073814, "grad_norm": 0.3732941150665283, "learning_rate": 8e-05, "loss": 1.5655, "step": 3954 }, { "epoch": 0.42742894196476816, "grad_norm": 0.3865974247455597, "learning_rate": 8e-05, "loss": 1.7048, "step": 3955 }, { "epoch": 0.42753701502215496, "grad_norm": 0.3838065564632416, "learning_rate": 8e-05, "loss": 1.7149, "step": 3956 }, { "epoch": 0.42764508807954177, "grad_norm": 0.3770924508571625, "learning_rate": 8e-05, "loss": 1.753, "step": 3957 }, { "epoch": 0.4277531611369286, "grad_norm": 0.38309067487716675, "learning_rate": 8e-05, "loss": 1.5574, "step": 3958 }, { "epoch": 0.4278612341943154, "grad_norm": 0.45394620299339294, "learning_rate": 8e-05, "loss": 1.7544, "step": 3959 }, { "epoch": 0.4279693072517021, "grad_norm": 0.4789636731147766, "learning_rate": 8e-05, "loss": 1.9358, "step": 3960 }, { "epoch": 0.42807738030908893, "grad_norm": 0.4569905400276184, "learning_rate": 8e-05, "loss": 1.9567, "step": 3961 }, { "epoch": 0.42818545336647573, "grad_norm": 0.3849007189273834, "learning_rate": 8e-05, "loss": 1.8783, "step": 3962 }, { "epoch": 0.42829352642386254, "grad_norm": 0.38390904664993286, "learning_rate": 8e-05, "loss": 1.7101, "step": 3963 }, { "epoch": 0.42840159948124934, "grad_norm": 0.3969222903251648, "learning_rate": 8e-05, "loss": 1.7045, "step": 3964 }, { "epoch": 0.42850967253863614, "grad_norm": 0.45306316018104553, "learning_rate": 8e-05, "loss": 1.7742, "step": 3965 }, { "epoch": 0.4286177455960229, "grad_norm": 0.3973192274570465, "learning_rate": 8e-05, "loss": 1.6415, "step": 3966 }, { "epoch": 0.4287258186534097, "grad_norm": 0.40596261620521545, "learning_rate": 8e-05, "loss": 1.7773, "step": 3967 }, { "epoch": 0.4288338917107965, "grad_norm": 0.4228540062904358, "learning_rate": 8e-05, "loss": 1.8254, "step": 3968 }, { "epoch": 0.4289419647681833, "grad_norm": 0.3980039358139038, "learning_rate": 8e-05, "loss": 1.8104, "step": 3969 }, { "epoch": 0.4290500378255701, "grad_norm": 0.4043446481227875, "learning_rate": 8e-05, "loss": 1.6738, "step": 3970 }, { "epoch": 0.42915811088295686, "grad_norm": 0.4232507348060608, "learning_rate": 8e-05, "loss": 1.8177, "step": 3971 }, { "epoch": 0.42926618394034366, "grad_norm": 0.49053773283958435, "learning_rate": 8e-05, "loss": 1.7424, "step": 3972 }, { "epoch": 0.42937425699773046, "grad_norm": 0.3948713541030884, "learning_rate": 8e-05, "loss": 1.7545, "step": 3973 }, { "epoch": 0.42948233005511727, "grad_norm": 0.3883204162120819, "learning_rate": 8e-05, "loss": 1.788, "step": 3974 }, { "epoch": 0.42959040311250407, "grad_norm": 0.38677725195884705, "learning_rate": 8e-05, "loss": 1.625, "step": 3975 }, { "epoch": 0.4296984761698908, "grad_norm": 0.4751206636428833, "learning_rate": 8e-05, "loss": 1.6998, "step": 3976 }, { "epoch": 0.4298065492272776, "grad_norm": 0.38869211077690125, "learning_rate": 8e-05, "loss": 1.6986, "step": 3977 }, { "epoch": 0.4299146222846644, "grad_norm": 0.49260348081588745, "learning_rate": 8e-05, "loss": 1.6929, "step": 3978 }, { "epoch": 0.43002269534205123, "grad_norm": 0.3923260569572449, "learning_rate": 8e-05, "loss": 1.7047, "step": 3979 }, { "epoch": 0.43013076839943803, "grad_norm": 0.3935205042362213, "learning_rate": 8e-05, "loss": 1.5993, "step": 3980 }, { "epoch": 0.43023884145682484, "grad_norm": 0.3812035918235779, "learning_rate": 8e-05, "loss": 1.5602, "step": 3981 }, { "epoch": 0.4303469145142116, "grad_norm": 0.37742292881011963, "learning_rate": 8e-05, "loss": 1.6258, "step": 3982 }, { "epoch": 0.4304549875715984, "grad_norm": 0.4100148677825928, "learning_rate": 8e-05, "loss": 1.7939, "step": 3983 }, { "epoch": 0.4305630606289852, "grad_norm": 0.42412251234054565, "learning_rate": 8e-05, "loss": 1.7596, "step": 3984 }, { "epoch": 0.430671133686372, "grad_norm": 0.38255825638771057, "learning_rate": 8e-05, "loss": 1.5164, "step": 3985 }, { "epoch": 0.4307792067437588, "grad_norm": 0.42709001898765564, "learning_rate": 8e-05, "loss": 1.7752, "step": 3986 }, { "epoch": 0.43088727980114555, "grad_norm": 0.3914938271045685, "learning_rate": 8e-05, "loss": 1.6601, "step": 3987 }, { "epoch": 0.43099535285853235, "grad_norm": 0.3754102289676666, "learning_rate": 8e-05, "loss": 1.7005, "step": 3988 }, { "epoch": 0.43110342591591916, "grad_norm": 0.41938459873199463, "learning_rate": 8e-05, "loss": 1.9189, "step": 3989 }, { "epoch": 0.43121149897330596, "grad_norm": 0.3966130018234253, "learning_rate": 8e-05, "loss": 1.8492, "step": 3990 }, { "epoch": 0.43131957203069277, "grad_norm": 0.3872619569301605, "learning_rate": 8e-05, "loss": 1.5972, "step": 3991 }, { "epoch": 0.4314276450880795, "grad_norm": 0.3895432949066162, "learning_rate": 8e-05, "loss": 1.6883, "step": 3992 }, { "epoch": 0.4315357181454663, "grad_norm": 0.4713430404663086, "learning_rate": 8e-05, "loss": 2.057, "step": 3993 }, { "epoch": 0.4316437912028531, "grad_norm": 0.4199851453304291, "learning_rate": 8e-05, "loss": 1.7573, "step": 3994 }, { "epoch": 0.4317518642602399, "grad_norm": 0.44189178943634033, "learning_rate": 8e-05, "loss": 1.7655, "step": 3995 }, { "epoch": 0.43185993731762673, "grad_norm": 0.4394003748893738, "learning_rate": 8e-05, "loss": 1.738, "step": 3996 }, { "epoch": 0.43196801037501353, "grad_norm": 0.3976317048072815, "learning_rate": 8e-05, "loss": 1.6999, "step": 3997 }, { "epoch": 0.4320760834324003, "grad_norm": 0.39891842007637024, "learning_rate": 8e-05, "loss": 1.6973, "step": 3998 }, { "epoch": 0.4321841564897871, "grad_norm": 0.4084338843822479, "learning_rate": 8e-05, "loss": 1.7959, "step": 3999 }, { "epoch": 0.4322922295471739, "grad_norm": 0.44293370842933655, "learning_rate": 8e-05, "loss": 1.7043, "step": 4000 }, { "epoch": 0.4324003026045607, "grad_norm": 0.4261878728866577, "learning_rate": 8e-05, "loss": 1.6225, "step": 4001 }, { "epoch": 0.4325083756619475, "grad_norm": 0.3853982985019684, "learning_rate": 8e-05, "loss": 1.6346, "step": 4002 }, { "epoch": 0.43261644871933425, "grad_norm": 0.3828014135360718, "learning_rate": 8e-05, "loss": 1.6691, "step": 4003 }, { "epoch": 0.43272452177672105, "grad_norm": 0.41485053300857544, "learning_rate": 8e-05, "loss": 1.6813, "step": 4004 }, { "epoch": 0.43283259483410785, "grad_norm": 0.4230251610279083, "learning_rate": 8e-05, "loss": 1.9448, "step": 4005 }, { "epoch": 0.43294066789149466, "grad_norm": 0.3784262239933014, "learning_rate": 8e-05, "loss": 1.6944, "step": 4006 }, { "epoch": 0.43304874094888146, "grad_norm": 0.38575780391693115, "learning_rate": 8e-05, "loss": 1.7421, "step": 4007 }, { "epoch": 0.43315681400626826, "grad_norm": 0.3846253752708435, "learning_rate": 8e-05, "loss": 1.4029, "step": 4008 }, { "epoch": 0.433264887063655, "grad_norm": 0.3932308256626129, "learning_rate": 8e-05, "loss": 1.6591, "step": 4009 }, { "epoch": 0.4333729601210418, "grad_norm": 0.4177047908306122, "learning_rate": 8e-05, "loss": 1.698, "step": 4010 }, { "epoch": 0.4334810331784286, "grad_norm": 0.4702097177505493, "learning_rate": 8e-05, "loss": 1.8476, "step": 4011 }, { "epoch": 0.4335891062358154, "grad_norm": 0.39412257075309753, "learning_rate": 8e-05, "loss": 1.6747, "step": 4012 }, { "epoch": 0.43369717929320223, "grad_norm": 0.3865263760089874, "learning_rate": 8e-05, "loss": 1.7784, "step": 4013 }, { "epoch": 0.433805252350589, "grad_norm": 0.42205196619033813, "learning_rate": 8e-05, "loss": 1.752, "step": 4014 }, { "epoch": 0.4339133254079758, "grad_norm": 0.4031464159488678, "learning_rate": 8e-05, "loss": 1.6739, "step": 4015 }, { "epoch": 0.4340213984653626, "grad_norm": 0.46766963601112366, "learning_rate": 8e-05, "loss": 1.9417, "step": 4016 }, { "epoch": 0.4341294715227494, "grad_norm": 0.4176258444786072, "learning_rate": 8e-05, "loss": 1.731, "step": 4017 }, { "epoch": 0.4342375445801362, "grad_norm": 0.415147066116333, "learning_rate": 8e-05, "loss": 1.6789, "step": 4018 }, { "epoch": 0.43434561763752294, "grad_norm": 0.3756083846092224, "learning_rate": 8e-05, "loss": 1.7074, "step": 4019 }, { "epoch": 0.43445369069490974, "grad_norm": 0.40239471197128296, "learning_rate": 8e-05, "loss": 1.7211, "step": 4020 }, { "epoch": 0.43456176375229655, "grad_norm": 0.3957957327365875, "learning_rate": 8e-05, "loss": 1.6966, "step": 4021 }, { "epoch": 0.43466983680968335, "grad_norm": 0.400188148021698, "learning_rate": 8e-05, "loss": 1.7349, "step": 4022 }, { "epoch": 0.43477790986707016, "grad_norm": 0.3612925708293915, "learning_rate": 8e-05, "loss": 1.5956, "step": 4023 }, { "epoch": 0.43488598292445696, "grad_norm": 0.3940187990665436, "learning_rate": 8e-05, "loss": 1.76, "step": 4024 }, { "epoch": 0.4349940559818437, "grad_norm": 0.44490760564804077, "learning_rate": 8e-05, "loss": 1.8895, "step": 4025 }, { "epoch": 0.4351021290392305, "grad_norm": 0.39821943640708923, "learning_rate": 8e-05, "loss": 1.8451, "step": 4026 }, { "epoch": 0.4352102020966173, "grad_norm": 0.3741059899330139, "learning_rate": 8e-05, "loss": 1.6333, "step": 4027 }, { "epoch": 0.4353182751540041, "grad_norm": 0.4242095649242401, "learning_rate": 8e-05, "loss": 1.8188, "step": 4028 }, { "epoch": 0.4354263482113909, "grad_norm": 0.4120519757270813, "learning_rate": 8e-05, "loss": 1.8413, "step": 4029 }, { "epoch": 0.43553442126877767, "grad_norm": 0.38116729259490967, "learning_rate": 8e-05, "loss": 1.772, "step": 4030 }, { "epoch": 0.4356424943261645, "grad_norm": 0.42187657952308655, "learning_rate": 8e-05, "loss": 1.85, "step": 4031 }, { "epoch": 0.4357505673835513, "grad_norm": 0.3659757375717163, "learning_rate": 8e-05, "loss": 1.6654, "step": 4032 }, { "epoch": 0.4358586404409381, "grad_norm": 0.42190250754356384, "learning_rate": 8e-05, "loss": 1.8069, "step": 4033 }, { "epoch": 0.4359667134983249, "grad_norm": 0.4002556800842285, "learning_rate": 8e-05, "loss": 1.782, "step": 4034 }, { "epoch": 0.43607478655571164, "grad_norm": 0.42683538794517517, "learning_rate": 8e-05, "loss": 1.9227, "step": 4035 }, { "epoch": 0.43618285961309844, "grad_norm": 0.37151455879211426, "learning_rate": 8e-05, "loss": 1.6294, "step": 4036 }, { "epoch": 0.43629093267048524, "grad_norm": 0.3944593369960785, "learning_rate": 8e-05, "loss": 1.7591, "step": 4037 }, { "epoch": 0.43639900572787205, "grad_norm": 0.37869954109191895, "learning_rate": 8e-05, "loss": 1.7272, "step": 4038 }, { "epoch": 0.43650707878525885, "grad_norm": 0.36323484778404236, "learning_rate": 8e-05, "loss": 1.6236, "step": 4039 }, { "epoch": 0.43661515184264565, "grad_norm": 0.387960821390152, "learning_rate": 8e-05, "loss": 1.6464, "step": 4040 }, { "epoch": 0.4367232249000324, "grad_norm": 0.44527754187583923, "learning_rate": 8e-05, "loss": 1.7654, "step": 4041 }, { "epoch": 0.4368312979574192, "grad_norm": 0.39944934844970703, "learning_rate": 8e-05, "loss": 1.6786, "step": 4042 }, { "epoch": 0.436939371014806, "grad_norm": 0.42589429020881653, "learning_rate": 8e-05, "loss": 1.5756, "step": 4043 }, { "epoch": 0.4370474440721928, "grad_norm": 0.42116838693618774, "learning_rate": 8e-05, "loss": 1.7177, "step": 4044 }, { "epoch": 0.4371555171295796, "grad_norm": 0.4094012379646301, "learning_rate": 8e-05, "loss": 1.7135, "step": 4045 }, { "epoch": 0.43726359018696637, "grad_norm": 0.39641955494880676, "learning_rate": 8e-05, "loss": 1.8137, "step": 4046 }, { "epoch": 0.43737166324435317, "grad_norm": 0.4124731421470642, "learning_rate": 8e-05, "loss": 1.9491, "step": 4047 }, { "epoch": 0.43747973630174, "grad_norm": 0.4227345287799835, "learning_rate": 8e-05, "loss": 1.8397, "step": 4048 }, { "epoch": 0.4375878093591268, "grad_norm": 0.38127344846725464, "learning_rate": 8e-05, "loss": 1.6297, "step": 4049 }, { "epoch": 0.4376958824165136, "grad_norm": 0.39730146527290344, "learning_rate": 8e-05, "loss": 1.6165, "step": 4050 }, { "epoch": 0.43780395547390033, "grad_norm": 0.39098402857780457, "learning_rate": 8e-05, "loss": 1.6508, "step": 4051 }, { "epoch": 0.43791202853128713, "grad_norm": 0.392983078956604, "learning_rate": 8e-05, "loss": 1.6921, "step": 4052 }, { "epoch": 0.43802010158867394, "grad_norm": 0.4320504069328308, "learning_rate": 8e-05, "loss": 1.8872, "step": 4053 }, { "epoch": 0.43812817464606074, "grad_norm": 0.44054827094078064, "learning_rate": 8e-05, "loss": 1.6491, "step": 4054 }, { "epoch": 0.43823624770344755, "grad_norm": 0.44570258259773254, "learning_rate": 8e-05, "loss": 1.9266, "step": 4055 }, { "epoch": 0.43834432076083435, "grad_norm": 0.40338799357414246, "learning_rate": 8e-05, "loss": 1.6142, "step": 4056 }, { "epoch": 0.4384523938182211, "grad_norm": 0.41716915369033813, "learning_rate": 8e-05, "loss": 1.7461, "step": 4057 }, { "epoch": 0.4385604668756079, "grad_norm": 0.40200549364089966, "learning_rate": 8e-05, "loss": 1.4431, "step": 4058 }, { "epoch": 0.4386685399329947, "grad_norm": 0.40294507145881653, "learning_rate": 8e-05, "loss": 1.649, "step": 4059 }, { "epoch": 0.4387766129903815, "grad_norm": 0.427957683801651, "learning_rate": 8e-05, "loss": 1.9719, "step": 4060 }, { "epoch": 0.4388846860477683, "grad_norm": 0.3961624503135681, "learning_rate": 8e-05, "loss": 1.6451, "step": 4061 }, { "epoch": 0.43899275910515506, "grad_norm": 0.40867164731025696, "learning_rate": 8e-05, "loss": 1.6683, "step": 4062 }, { "epoch": 0.43910083216254187, "grad_norm": 0.4313106834888458, "learning_rate": 8e-05, "loss": 1.8487, "step": 4063 }, { "epoch": 0.43920890521992867, "grad_norm": 0.380979061126709, "learning_rate": 8e-05, "loss": 1.7664, "step": 4064 }, { "epoch": 0.4393169782773155, "grad_norm": 0.4122730791568756, "learning_rate": 8e-05, "loss": 1.6375, "step": 4065 }, { "epoch": 0.4394250513347023, "grad_norm": 0.3955979645252228, "learning_rate": 8e-05, "loss": 1.778, "step": 4066 }, { "epoch": 0.439533124392089, "grad_norm": 0.401302695274353, "learning_rate": 8e-05, "loss": 1.8593, "step": 4067 }, { "epoch": 0.43964119744947583, "grad_norm": 0.385590136051178, "learning_rate": 8e-05, "loss": 1.7359, "step": 4068 }, { "epoch": 0.43974927050686263, "grad_norm": 0.4865040183067322, "learning_rate": 8e-05, "loss": 1.8133, "step": 4069 }, { "epoch": 0.43985734356424944, "grad_norm": 0.42025017738342285, "learning_rate": 8e-05, "loss": 1.7527, "step": 4070 }, { "epoch": 0.43996541662163624, "grad_norm": 0.39710304141044617, "learning_rate": 8e-05, "loss": 1.7476, "step": 4071 }, { "epoch": 0.44007348967902304, "grad_norm": 0.37739142775535583, "learning_rate": 8e-05, "loss": 1.692, "step": 4072 }, { "epoch": 0.4401815627364098, "grad_norm": 0.4381338655948639, "learning_rate": 8e-05, "loss": 1.8339, "step": 4073 }, { "epoch": 0.4402896357937966, "grad_norm": 0.4066935181617737, "learning_rate": 8e-05, "loss": 1.8268, "step": 4074 }, { "epoch": 0.4403977088511834, "grad_norm": 0.39243197441101074, "learning_rate": 8e-05, "loss": 1.7776, "step": 4075 }, { "epoch": 0.4405057819085702, "grad_norm": 0.3946973383426666, "learning_rate": 8e-05, "loss": 1.7883, "step": 4076 }, { "epoch": 0.440613854965957, "grad_norm": 0.37950167059898376, "learning_rate": 8e-05, "loss": 1.6927, "step": 4077 }, { "epoch": 0.44072192802334376, "grad_norm": 0.3903473913669586, "learning_rate": 8e-05, "loss": 1.7724, "step": 4078 }, { "epoch": 0.44083000108073056, "grad_norm": 0.3659137487411499, "learning_rate": 8e-05, "loss": 1.6077, "step": 4079 }, { "epoch": 0.44093807413811736, "grad_norm": 0.36249393224716187, "learning_rate": 8e-05, "loss": 1.6665, "step": 4080 }, { "epoch": 0.44104614719550417, "grad_norm": 0.38518384099006653, "learning_rate": 8e-05, "loss": 1.6907, "step": 4081 }, { "epoch": 0.44115422025289097, "grad_norm": 0.3802357316017151, "learning_rate": 8e-05, "loss": 1.7006, "step": 4082 }, { "epoch": 0.4412622933102777, "grad_norm": 0.4164290726184845, "learning_rate": 8e-05, "loss": 1.7803, "step": 4083 }, { "epoch": 0.4413703663676645, "grad_norm": 0.3685266375541687, "learning_rate": 8e-05, "loss": 1.5313, "step": 4084 }, { "epoch": 0.4414784394250513, "grad_norm": 0.3983461558818817, "learning_rate": 8e-05, "loss": 1.7748, "step": 4085 }, { "epoch": 0.44158651248243813, "grad_norm": 0.38751694560050964, "learning_rate": 8e-05, "loss": 1.5709, "step": 4086 }, { "epoch": 0.44169458553982494, "grad_norm": 0.418019562959671, "learning_rate": 8e-05, "loss": 1.8252, "step": 4087 }, { "epoch": 0.44180265859721174, "grad_norm": 0.42111557722091675, "learning_rate": 8e-05, "loss": 1.93, "step": 4088 }, { "epoch": 0.4419107316545985, "grad_norm": 0.3989101052284241, "learning_rate": 8e-05, "loss": 1.7258, "step": 4089 }, { "epoch": 0.4420188047119853, "grad_norm": 0.3857951760292053, "learning_rate": 8e-05, "loss": 1.6192, "step": 4090 }, { "epoch": 0.4421268777693721, "grad_norm": 0.39393389225006104, "learning_rate": 8e-05, "loss": 1.7811, "step": 4091 }, { "epoch": 0.4422349508267589, "grad_norm": 0.4382179081439972, "learning_rate": 8e-05, "loss": 1.765, "step": 4092 }, { "epoch": 0.4423430238841457, "grad_norm": 0.4087485074996948, "learning_rate": 8e-05, "loss": 1.6377, "step": 4093 }, { "epoch": 0.44245109694153245, "grad_norm": 0.3930310308933258, "learning_rate": 8e-05, "loss": 1.7044, "step": 4094 }, { "epoch": 0.44255916999891926, "grad_norm": 0.3805514872074127, "learning_rate": 8e-05, "loss": 1.7275, "step": 4095 }, { "epoch": 0.44266724305630606, "grad_norm": 0.4075923264026642, "learning_rate": 8e-05, "loss": 1.7615, "step": 4096 }, { "epoch": 0.44277531611369286, "grad_norm": 0.38746529817581177, "learning_rate": 8e-05, "loss": 1.675, "step": 4097 }, { "epoch": 0.44288338917107967, "grad_norm": 0.3932340443134308, "learning_rate": 8e-05, "loss": 1.6856, "step": 4098 }, { "epoch": 0.44299146222846647, "grad_norm": 0.3965499699115753, "learning_rate": 8e-05, "loss": 1.6854, "step": 4099 }, { "epoch": 0.4430995352858532, "grad_norm": 0.4528593122959137, "learning_rate": 8e-05, "loss": 1.9614, "step": 4100 }, { "epoch": 0.44320760834324, "grad_norm": 0.38144510984420776, "learning_rate": 8e-05, "loss": 1.7119, "step": 4101 }, { "epoch": 0.4433156814006268, "grad_norm": 0.3901524841785431, "learning_rate": 8e-05, "loss": 1.7622, "step": 4102 }, { "epoch": 0.44342375445801363, "grad_norm": 0.4248017966747284, "learning_rate": 8e-05, "loss": 1.8402, "step": 4103 }, { "epoch": 0.44353182751540043, "grad_norm": 0.4670308828353882, "learning_rate": 8e-05, "loss": 1.772, "step": 4104 }, { "epoch": 0.4436399005727872, "grad_norm": 0.4132468104362488, "learning_rate": 8e-05, "loss": 1.8095, "step": 4105 }, { "epoch": 0.443747973630174, "grad_norm": 0.38740426301956177, "learning_rate": 8e-05, "loss": 1.8122, "step": 4106 }, { "epoch": 0.4438560466875608, "grad_norm": 0.35123327374458313, "learning_rate": 8e-05, "loss": 1.5147, "step": 4107 }, { "epoch": 0.4439641197449476, "grad_norm": 0.360033243894577, "learning_rate": 8e-05, "loss": 1.5321, "step": 4108 }, { "epoch": 0.4440721928023344, "grad_norm": 0.3886578381061554, "learning_rate": 8e-05, "loss": 1.7402, "step": 4109 }, { "epoch": 0.44418026585972115, "grad_norm": 0.38930684328079224, "learning_rate": 8e-05, "loss": 1.8175, "step": 4110 }, { "epoch": 0.44428833891710795, "grad_norm": 0.38217923045158386, "learning_rate": 8e-05, "loss": 1.4983, "step": 4111 }, { "epoch": 0.44439641197449475, "grad_norm": 0.45865002274513245, "learning_rate": 8e-05, "loss": 1.7772, "step": 4112 }, { "epoch": 0.44450448503188156, "grad_norm": 0.45548492670059204, "learning_rate": 8e-05, "loss": 1.8117, "step": 4113 }, { "epoch": 0.44461255808926836, "grad_norm": 0.4268307089805603, "learning_rate": 8e-05, "loss": 2.0046, "step": 4114 }, { "epoch": 0.44472063114665517, "grad_norm": 0.41946035623550415, "learning_rate": 8e-05, "loss": 1.863, "step": 4115 }, { "epoch": 0.4448287042040419, "grad_norm": 0.43050646781921387, "learning_rate": 8e-05, "loss": 1.8613, "step": 4116 }, { "epoch": 0.4449367772614287, "grad_norm": 0.38218268752098083, "learning_rate": 8e-05, "loss": 1.6927, "step": 4117 }, { "epoch": 0.4450448503188155, "grad_norm": 0.3812640607357025, "learning_rate": 8e-05, "loss": 1.7996, "step": 4118 }, { "epoch": 0.4451529233762023, "grad_norm": 0.4203759729862213, "learning_rate": 8e-05, "loss": 1.8412, "step": 4119 }, { "epoch": 0.44526099643358913, "grad_norm": 0.39401933550834656, "learning_rate": 8e-05, "loss": 1.6846, "step": 4120 }, { "epoch": 0.4453690694909759, "grad_norm": 0.41144585609436035, "learning_rate": 8e-05, "loss": 1.7655, "step": 4121 }, { "epoch": 0.4454771425483627, "grad_norm": 0.40371081233024597, "learning_rate": 8e-05, "loss": 1.7963, "step": 4122 }, { "epoch": 0.4455852156057495, "grad_norm": 0.38249272108078003, "learning_rate": 8e-05, "loss": 1.7287, "step": 4123 }, { "epoch": 0.4456932886631363, "grad_norm": 0.40065309405326843, "learning_rate": 8e-05, "loss": 1.6287, "step": 4124 }, { "epoch": 0.4458013617205231, "grad_norm": 0.39524564146995544, "learning_rate": 8e-05, "loss": 1.8478, "step": 4125 }, { "epoch": 0.44590943477790984, "grad_norm": 0.40700867772102356, "learning_rate": 8e-05, "loss": 1.7352, "step": 4126 }, { "epoch": 0.44601750783529664, "grad_norm": 0.40793606638908386, "learning_rate": 8e-05, "loss": 1.8052, "step": 4127 }, { "epoch": 0.44612558089268345, "grad_norm": 0.43469879031181335, "learning_rate": 8e-05, "loss": 1.7806, "step": 4128 }, { "epoch": 0.44623365395007025, "grad_norm": 0.40545299649238586, "learning_rate": 8e-05, "loss": 1.7583, "step": 4129 }, { "epoch": 0.44634172700745706, "grad_norm": 0.4154534935951233, "learning_rate": 8e-05, "loss": 1.8038, "step": 4130 }, { "epoch": 0.44644980006484386, "grad_norm": 0.37462887167930603, "learning_rate": 8e-05, "loss": 1.6905, "step": 4131 }, { "epoch": 0.4465578731222306, "grad_norm": 0.4171428680419922, "learning_rate": 8e-05, "loss": 1.7983, "step": 4132 }, { "epoch": 0.4466659461796174, "grad_norm": 0.40016424655914307, "learning_rate": 8e-05, "loss": 1.7067, "step": 4133 }, { "epoch": 0.4467740192370042, "grad_norm": 0.4300216734409332, "learning_rate": 8e-05, "loss": 1.8315, "step": 4134 }, { "epoch": 0.446882092294391, "grad_norm": 0.4333845376968384, "learning_rate": 8e-05, "loss": 1.8441, "step": 4135 }, { "epoch": 0.4469901653517778, "grad_norm": 0.38947948813438416, "learning_rate": 8e-05, "loss": 1.5957, "step": 4136 }, { "epoch": 0.4470982384091646, "grad_norm": 0.3804239332675934, "learning_rate": 8e-05, "loss": 1.6908, "step": 4137 }, { "epoch": 0.4472063114665514, "grad_norm": 0.39848506450653076, "learning_rate": 8e-05, "loss": 1.7669, "step": 4138 }, { "epoch": 0.4473143845239382, "grad_norm": 0.4335612952709198, "learning_rate": 8e-05, "loss": 1.6697, "step": 4139 }, { "epoch": 0.447422457581325, "grad_norm": 0.395724892616272, "learning_rate": 8e-05, "loss": 1.6947, "step": 4140 }, { "epoch": 0.4475305306387118, "grad_norm": 0.4212934672832489, "learning_rate": 8e-05, "loss": 1.7986, "step": 4141 }, { "epoch": 0.44763860369609854, "grad_norm": 0.3814888000488281, "learning_rate": 8e-05, "loss": 1.563, "step": 4142 }, { "epoch": 0.44774667675348534, "grad_norm": 0.3882804811000824, "learning_rate": 8e-05, "loss": 1.6502, "step": 4143 }, { "epoch": 0.44785474981087214, "grad_norm": 0.409474641084671, "learning_rate": 8e-05, "loss": 1.6091, "step": 4144 }, { "epoch": 0.44796282286825895, "grad_norm": 0.36955225467681885, "learning_rate": 8e-05, "loss": 1.6342, "step": 4145 }, { "epoch": 0.44807089592564575, "grad_norm": 0.3800424635410309, "learning_rate": 8e-05, "loss": 1.6154, "step": 4146 }, { "epoch": 0.44817896898303256, "grad_norm": 0.41920849680900574, "learning_rate": 8e-05, "loss": 1.631, "step": 4147 }, { "epoch": 0.4482870420404193, "grad_norm": 0.4319944381713867, "learning_rate": 8e-05, "loss": 1.6814, "step": 4148 }, { "epoch": 0.4483951150978061, "grad_norm": 0.4809088706970215, "learning_rate": 8e-05, "loss": 1.9423, "step": 4149 }, { "epoch": 0.4485031881551929, "grad_norm": 0.4052334427833557, "learning_rate": 8e-05, "loss": 1.7557, "step": 4150 }, { "epoch": 0.4486112612125797, "grad_norm": 0.3898913562297821, "learning_rate": 8e-05, "loss": 1.6813, "step": 4151 }, { "epoch": 0.4487193342699665, "grad_norm": 0.41606011986732483, "learning_rate": 8e-05, "loss": 1.8142, "step": 4152 }, { "epoch": 0.44882740732735327, "grad_norm": 0.38073471188545227, "learning_rate": 8e-05, "loss": 1.7149, "step": 4153 }, { "epoch": 0.44893548038474007, "grad_norm": 0.41051143407821655, "learning_rate": 8e-05, "loss": 1.5759, "step": 4154 }, { "epoch": 0.4490435534421269, "grad_norm": 0.39827659726142883, "learning_rate": 8e-05, "loss": 1.7553, "step": 4155 }, { "epoch": 0.4491516264995137, "grad_norm": 0.39858707785606384, "learning_rate": 8e-05, "loss": 1.6401, "step": 4156 }, { "epoch": 0.4492596995569005, "grad_norm": 0.41112565994262695, "learning_rate": 8e-05, "loss": 1.7226, "step": 4157 }, { "epoch": 0.44936777261428723, "grad_norm": 0.38678601384162903, "learning_rate": 8e-05, "loss": 1.7793, "step": 4158 }, { "epoch": 0.44947584567167403, "grad_norm": 0.3781753182411194, "learning_rate": 8e-05, "loss": 1.6676, "step": 4159 }, { "epoch": 0.44958391872906084, "grad_norm": 0.39570656418800354, "learning_rate": 8e-05, "loss": 1.7075, "step": 4160 }, { "epoch": 0.44969199178644764, "grad_norm": 0.38605526089668274, "learning_rate": 8e-05, "loss": 1.6397, "step": 4161 }, { "epoch": 0.44980006484383445, "grad_norm": 0.3786720931529999, "learning_rate": 8e-05, "loss": 1.6792, "step": 4162 }, { "epoch": 0.44990813790122125, "grad_norm": 0.39362895488739014, "learning_rate": 8e-05, "loss": 1.4707, "step": 4163 }, { "epoch": 0.450016210958608, "grad_norm": 0.3991580009460449, "learning_rate": 8e-05, "loss": 1.7009, "step": 4164 }, { "epoch": 0.4501242840159948, "grad_norm": 0.3815881311893463, "learning_rate": 8e-05, "loss": 1.6125, "step": 4165 }, { "epoch": 0.4502323570733816, "grad_norm": 0.4243254065513611, "learning_rate": 8e-05, "loss": 1.8755, "step": 4166 }, { "epoch": 0.4503404301307684, "grad_norm": 0.37091976404190063, "learning_rate": 8e-05, "loss": 1.536, "step": 4167 }, { "epoch": 0.4504485031881552, "grad_norm": 0.3933130204677582, "learning_rate": 8e-05, "loss": 1.7422, "step": 4168 }, { "epoch": 0.45055657624554196, "grad_norm": 0.41969093680381775, "learning_rate": 8e-05, "loss": 1.8135, "step": 4169 }, { "epoch": 0.45066464930292877, "grad_norm": 0.40764015913009644, "learning_rate": 8e-05, "loss": 1.6265, "step": 4170 }, { "epoch": 0.45077272236031557, "grad_norm": 0.39879798889160156, "learning_rate": 8e-05, "loss": 1.7237, "step": 4171 }, { "epoch": 0.4508807954177024, "grad_norm": 0.4008561074733734, "learning_rate": 8e-05, "loss": 1.7057, "step": 4172 }, { "epoch": 0.4509888684750892, "grad_norm": 0.3942168354988098, "learning_rate": 8e-05, "loss": 1.866, "step": 4173 }, { "epoch": 0.451096941532476, "grad_norm": 0.4003502428531647, "learning_rate": 8e-05, "loss": 1.7716, "step": 4174 }, { "epoch": 0.45120501458986273, "grad_norm": 0.38686972856521606, "learning_rate": 8e-05, "loss": 1.7233, "step": 4175 }, { "epoch": 0.45131308764724953, "grad_norm": 0.3975874185562134, "learning_rate": 8e-05, "loss": 1.6512, "step": 4176 }, { "epoch": 0.45142116070463634, "grad_norm": 0.4099818170070648, "learning_rate": 8e-05, "loss": 1.6666, "step": 4177 }, { "epoch": 0.45152923376202314, "grad_norm": 0.39071857929229736, "learning_rate": 8e-05, "loss": 1.8683, "step": 4178 }, { "epoch": 0.45163730681940994, "grad_norm": 0.40196681022644043, "learning_rate": 8e-05, "loss": 1.5938, "step": 4179 }, { "epoch": 0.4517453798767967, "grad_norm": 0.39036962389945984, "learning_rate": 8e-05, "loss": 1.7004, "step": 4180 }, { "epoch": 0.4518534529341835, "grad_norm": 0.38082221150398254, "learning_rate": 8e-05, "loss": 1.7278, "step": 4181 }, { "epoch": 0.4519615259915703, "grad_norm": 0.40840616822242737, "learning_rate": 8e-05, "loss": 1.5474, "step": 4182 }, { "epoch": 0.4520695990489571, "grad_norm": 0.366936594247818, "learning_rate": 8e-05, "loss": 1.598, "step": 4183 }, { "epoch": 0.4521776721063439, "grad_norm": 0.43335720896720886, "learning_rate": 8e-05, "loss": 1.8735, "step": 4184 }, { "epoch": 0.45228574516373066, "grad_norm": 0.4122973084449768, "learning_rate": 8e-05, "loss": 1.7242, "step": 4185 }, { "epoch": 0.45239381822111746, "grad_norm": 0.38433751463890076, "learning_rate": 8e-05, "loss": 1.5977, "step": 4186 }, { "epoch": 0.45250189127850426, "grad_norm": 0.3955709934234619, "learning_rate": 8e-05, "loss": 1.8994, "step": 4187 }, { "epoch": 0.45260996433589107, "grad_norm": 0.38264644145965576, "learning_rate": 8e-05, "loss": 1.6156, "step": 4188 }, { "epoch": 0.4527180373932779, "grad_norm": 0.4050297141075134, "learning_rate": 8e-05, "loss": 1.6352, "step": 4189 }, { "epoch": 0.4528261104506647, "grad_norm": 0.3812471628189087, "learning_rate": 8e-05, "loss": 1.6091, "step": 4190 }, { "epoch": 0.4529341835080514, "grad_norm": 0.3948255181312561, "learning_rate": 8e-05, "loss": 1.7144, "step": 4191 }, { "epoch": 0.45304225656543823, "grad_norm": 0.4553428292274475, "learning_rate": 8e-05, "loss": 1.7806, "step": 4192 }, { "epoch": 0.45315032962282503, "grad_norm": 0.42498478293418884, "learning_rate": 8e-05, "loss": 1.7588, "step": 4193 }, { "epoch": 0.45325840268021184, "grad_norm": 0.3939964473247528, "learning_rate": 8e-05, "loss": 1.8188, "step": 4194 }, { "epoch": 0.45336647573759864, "grad_norm": 0.3810969889163971, "learning_rate": 8e-05, "loss": 1.7752, "step": 4195 }, { "epoch": 0.4534745487949854, "grad_norm": 0.42184048891067505, "learning_rate": 8e-05, "loss": 1.8841, "step": 4196 }, { "epoch": 0.4535826218523722, "grad_norm": 0.41052091121673584, "learning_rate": 8e-05, "loss": 1.6255, "step": 4197 }, { "epoch": 0.453690694909759, "grad_norm": 0.4232829511165619, "learning_rate": 8e-05, "loss": 1.7828, "step": 4198 }, { "epoch": 0.4537987679671458, "grad_norm": 0.3979233503341675, "learning_rate": 8e-05, "loss": 1.7539, "step": 4199 }, { "epoch": 0.4539068410245326, "grad_norm": 0.4140111804008484, "learning_rate": 8e-05, "loss": 1.7158, "step": 4200 }, { "epoch": 0.45401491408191935, "grad_norm": 0.4476223289966583, "learning_rate": 8e-05, "loss": 1.8369, "step": 4201 }, { "epoch": 0.45412298713930616, "grad_norm": 0.4117853343486786, "learning_rate": 8e-05, "loss": 1.6882, "step": 4202 }, { "epoch": 0.45423106019669296, "grad_norm": 0.38173508644104004, "learning_rate": 8e-05, "loss": 1.7465, "step": 4203 }, { "epoch": 0.45433913325407976, "grad_norm": 0.41248437762260437, "learning_rate": 8e-05, "loss": 1.6825, "step": 4204 }, { "epoch": 0.45444720631146657, "grad_norm": 0.4649130702018738, "learning_rate": 8e-05, "loss": 1.7466, "step": 4205 }, { "epoch": 0.45455527936885337, "grad_norm": 0.4088135361671448, "learning_rate": 8e-05, "loss": 1.8455, "step": 4206 }, { "epoch": 0.4546633524262401, "grad_norm": 0.3916531801223755, "learning_rate": 8e-05, "loss": 1.5961, "step": 4207 }, { "epoch": 0.4547714254836269, "grad_norm": 0.3791961073875427, "learning_rate": 8e-05, "loss": 1.6748, "step": 4208 }, { "epoch": 0.4548794985410137, "grad_norm": 0.40511149168014526, "learning_rate": 8e-05, "loss": 1.7729, "step": 4209 }, { "epoch": 0.45498757159840053, "grad_norm": 0.4018228054046631, "learning_rate": 8e-05, "loss": 1.8085, "step": 4210 }, { "epoch": 0.45509564465578733, "grad_norm": 0.3974384069442749, "learning_rate": 8e-05, "loss": 1.7313, "step": 4211 }, { "epoch": 0.4552037177131741, "grad_norm": 0.3767494857311249, "learning_rate": 8e-05, "loss": 1.7715, "step": 4212 }, { "epoch": 0.4553117907705609, "grad_norm": 0.43515220284461975, "learning_rate": 8e-05, "loss": 1.6038, "step": 4213 }, { "epoch": 0.4554198638279477, "grad_norm": 0.3716447353363037, "learning_rate": 8e-05, "loss": 1.6678, "step": 4214 }, { "epoch": 0.4555279368853345, "grad_norm": 0.36889275908470154, "learning_rate": 8e-05, "loss": 1.6402, "step": 4215 }, { "epoch": 0.4556360099427213, "grad_norm": 0.3953246772289276, "learning_rate": 8e-05, "loss": 1.8373, "step": 4216 }, { "epoch": 0.45574408300010805, "grad_norm": 0.40527579188346863, "learning_rate": 8e-05, "loss": 1.6154, "step": 4217 }, { "epoch": 0.45585215605749485, "grad_norm": 0.40630364418029785, "learning_rate": 8e-05, "loss": 1.8841, "step": 4218 }, { "epoch": 0.45596022911488165, "grad_norm": 0.3697873055934906, "learning_rate": 8e-05, "loss": 1.7918, "step": 4219 }, { "epoch": 0.45606830217226846, "grad_norm": 0.3869011104106903, "learning_rate": 8e-05, "loss": 1.7002, "step": 4220 }, { "epoch": 0.45617637522965526, "grad_norm": 0.4355603754520416, "learning_rate": 8e-05, "loss": 1.8407, "step": 4221 }, { "epoch": 0.45628444828704207, "grad_norm": 0.4080151319503784, "learning_rate": 8e-05, "loss": 1.832, "step": 4222 }, { "epoch": 0.4563925213444288, "grad_norm": 0.37484556436538696, "learning_rate": 8e-05, "loss": 1.7048, "step": 4223 }, { "epoch": 0.4565005944018156, "grad_norm": 0.38602215051651, "learning_rate": 8e-05, "loss": 1.7111, "step": 4224 }, { "epoch": 0.4566086674592024, "grad_norm": 0.41468873620033264, "learning_rate": 8e-05, "loss": 1.8882, "step": 4225 }, { "epoch": 0.4567167405165892, "grad_norm": 0.40375640988349915, "learning_rate": 8e-05, "loss": 1.8361, "step": 4226 }, { "epoch": 0.45682481357397603, "grad_norm": 0.39360880851745605, "learning_rate": 8e-05, "loss": 1.6172, "step": 4227 }, { "epoch": 0.4569328866313628, "grad_norm": 0.4090273380279541, "learning_rate": 8e-05, "loss": 1.5416, "step": 4228 }, { "epoch": 0.4570409596887496, "grad_norm": 0.41469478607177734, "learning_rate": 8e-05, "loss": 1.6082, "step": 4229 }, { "epoch": 0.4571490327461364, "grad_norm": 0.40205758810043335, "learning_rate": 8e-05, "loss": 1.7276, "step": 4230 }, { "epoch": 0.4572571058035232, "grad_norm": 0.37132859230041504, "learning_rate": 8e-05, "loss": 1.6877, "step": 4231 }, { "epoch": 0.45736517886091, "grad_norm": 0.38051095604896545, "learning_rate": 8e-05, "loss": 1.7735, "step": 4232 }, { "epoch": 0.45747325191829674, "grad_norm": 0.3900478780269623, "learning_rate": 8e-05, "loss": 1.712, "step": 4233 }, { "epoch": 0.45758132497568355, "grad_norm": 0.3997434973716736, "learning_rate": 8e-05, "loss": 1.7378, "step": 4234 }, { "epoch": 0.45768939803307035, "grad_norm": 0.3884037137031555, "learning_rate": 8e-05, "loss": 1.4314, "step": 4235 }, { "epoch": 0.45779747109045715, "grad_norm": 0.38705649971961975, "learning_rate": 8e-05, "loss": 1.7249, "step": 4236 }, { "epoch": 0.45790554414784396, "grad_norm": 0.3733726143836975, "learning_rate": 8e-05, "loss": 1.4933, "step": 4237 }, { "epoch": 0.45801361720523076, "grad_norm": 0.4090176224708557, "learning_rate": 8e-05, "loss": 1.7865, "step": 4238 }, { "epoch": 0.4581216902626175, "grad_norm": 0.3869929909706116, "learning_rate": 8e-05, "loss": 1.5816, "step": 4239 }, { "epoch": 0.4582297633200043, "grad_norm": 0.402170866727829, "learning_rate": 8e-05, "loss": 1.6822, "step": 4240 }, { "epoch": 0.4583378363773911, "grad_norm": 0.3807542026042938, "learning_rate": 8e-05, "loss": 1.6426, "step": 4241 }, { "epoch": 0.4584459094347779, "grad_norm": 0.4597344696521759, "learning_rate": 8e-05, "loss": 1.6257, "step": 4242 }, { "epoch": 0.4585539824921647, "grad_norm": 0.48820099234580994, "learning_rate": 8e-05, "loss": 1.9257, "step": 4243 }, { "epoch": 0.4586620555495515, "grad_norm": 0.3972399830818176, "learning_rate": 8e-05, "loss": 1.6972, "step": 4244 }, { "epoch": 0.4587701286069383, "grad_norm": 0.4314461946487427, "learning_rate": 8e-05, "loss": 1.8162, "step": 4245 }, { "epoch": 0.4588782016643251, "grad_norm": 0.40236297249794006, "learning_rate": 8e-05, "loss": 1.7352, "step": 4246 }, { "epoch": 0.4589862747217119, "grad_norm": 0.384321928024292, "learning_rate": 8e-05, "loss": 1.7116, "step": 4247 }, { "epoch": 0.4590943477790987, "grad_norm": 0.41385918855667114, "learning_rate": 8e-05, "loss": 1.4514, "step": 4248 }, { "epoch": 0.45920242083648544, "grad_norm": 0.45319339632987976, "learning_rate": 8e-05, "loss": 1.8928, "step": 4249 }, { "epoch": 0.45931049389387224, "grad_norm": 0.3914586305618286, "learning_rate": 8e-05, "loss": 1.7159, "step": 4250 }, { "epoch": 0.45941856695125904, "grad_norm": 0.39298906922340393, "learning_rate": 8e-05, "loss": 1.6449, "step": 4251 }, { "epoch": 0.45952664000864585, "grad_norm": 0.3825382888317108, "learning_rate": 8e-05, "loss": 1.6443, "step": 4252 }, { "epoch": 0.45963471306603265, "grad_norm": 0.39020484685897827, "learning_rate": 8e-05, "loss": 1.731, "step": 4253 }, { "epoch": 0.45974278612341946, "grad_norm": 0.3996971547603607, "learning_rate": 8e-05, "loss": 1.6702, "step": 4254 }, { "epoch": 0.4598508591808062, "grad_norm": 0.4055745601654053, "learning_rate": 8e-05, "loss": 1.8496, "step": 4255 }, { "epoch": 0.459958932238193, "grad_norm": 0.3965752422809601, "learning_rate": 8e-05, "loss": 1.6494, "step": 4256 }, { "epoch": 0.4600670052955798, "grad_norm": 0.39176952838897705, "learning_rate": 8e-05, "loss": 1.7194, "step": 4257 }, { "epoch": 0.4601750783529666, "grad_norm": 0.394587904214859, "learning_rate": 8e-05, "loss": 1.714, "step": 4258 }, { "epoch": 0.4602831514103534, "grad_norm": 0.38011595606803894, "learning_rate": 8e-05, "loss": 1.6683, "step": 4259 }, { "epoch": 0.46039122446774017, "grad_norm": 0.3950192928314209, "learning_rate": 8e-05, "loss": 1.7885, "step": 4260 }, { "epoch": 0.46049929752512697, "grad_norm": 0.4274528920650482, "learning_rate": 8e-05, "loss": 1.8044, "step": 4261 }, { "epoch": 0.4606073705825138, "grad_norm": 0.4229106903076172, "learning_rate": 8e-05, "loss": 1.7517, "step": 4262 }, { "epoch": 0.4607154436399006, "grad_norm": 0.3834282457828522, "learning_rate": 8e-05, "loss": 1.6904, "step": 4263 }, { "epoch": 0.4608235166972874, "grad_norm": 0.3826853930950165, "learning_rate": 8e-05, "loss": 1.555, "step": 4264 }, { "epoch": 0.4609315897546742, "grad_norm": 0.35426029562950134, "learning_rate": 8e-05, "loss": 1.5482, "step": 4265 }, { "epoch": 0.46103966281206094, "grad_norm": 0.4941841661930084, "learning_rate": 8e-05, "loss": 2.1103, "step": 4266 }, { "epoch": 0.46114773586944774, "grad_norm": 0.38645580410957336, "learning_rate": 8e-05, "loss": 1.6807, "step": 4267 }, { "epoch": 0.46125580892683454, "grad_norm": 0.3816952407360077, "learning_rate": 8e-05, "loss": 1.6676, "step": 4268 }, { "epoch": 0.46136388198422135, "grad_norm": 0.3699115514755249, "learning_rate": 8e-05, "loss": 1.7383, "step": 4269 }, { "epoch": 0.46147195504160815, "grad_norm": 0.4379904866218567, "learning_rate": 8e-05, "loss": 1.7083, "step": 4270 }, { "epoch": 0.4615800280989949, "grad_norm": 0.40712636709213257, "learning_rate": 8e-05, "loss": 1.7554, "step": 4271 }, { "epoch": 0.4616881011563817, "grad_norm": 0.42166948318481445, "learning_rate": 8e-05, "loss": 1.727, "step": 4272 }, { "epoch": 0.4617961742137685, "grad_norm": 0.4893457889556885, "learning_rate": 8e-05, "loss": 1.8257, "step": 4273 }, { "epoch": 0.4619042472711553, "grad_norm": 0.3916114270687103, "learning_rate": 8e-05, "loss": 1.7031, "step": 4274 }, { "epoch": 0.4620123203285421, "grad_norm": 0.41695448756217957, "learning_rate": 8e-05, "loss": 1.7869, "step": 4275 }, { "epoch": 0.46212039338592886, "grad_norm": 0.38892853260040283, "learning_rate": 8e-05, "loss": 1.6805, "step": 4276 }, { "epoch": 0.46222846644331567, "grad_norm": 0.40539973974227905, "learning_rate": 8e-05, "loss": 1.6894, "step": 4277 }, { "epoch": 0.46233653950070247, "grad_norm": 0.4102711081504822, "learning_rate": 8e-05, "loss": 1.8844, "step": 4278 }, { "epoch": 0.4624446125580893, "grad_norm": 0.38708868622779846, "learning_rate": 8e-05, "loss": 1.6971, "step": 4279 }, { "epoch": 0.4625526856154761, "grad_norm": 0.39683011174201965, "learning_rate": 8e-05, "loss": 1.5848, "step": 4280 }, { "epoch": 0.4626607586728629, "grad_norm": 0.4241722524166107, "learning_rate": 8e-05, "loss": 1.87, "step": 4281 }, { "epoch": 0.46276883173024963, "grad_norm": 0.4027661383152008, "learning_rate": 8e-05, "loss": 1.8335, "step": 4282 }, { "epoch": 0.46287690478763643, "grad_norm": 0.3885311782360077, "learning_rate": 8e-05, "loss": 1.7674, "step": 4283 }, { "epoch": 0.46298497784502324, "grad_norm": 0.3851543068885803, "learning_rate": 8e-05, "loss": 1.6956, "step": 4284 }, { "epoch": 0.46309305090241004, "grad_norm": 0.3999451994895935, "learning_rate": 8e-05, "loss": 1.7463, "step": 4285 }, { "epoch": 0.46320112395979685, "grad_norm": 0.406423956155777, "learning_rate": 8e-05, "loss": 1.8638, "step": 4286 }, { "epoch": 0.4633091970171836, "grad_norm": 0.3860330879688263, "learning_rate": 8e-05, "loss": 1.7611, "step": 4287 }, { "epoch": 0.4634172700745704, "grad_norm": 0.390470027923584, "learning_rate": 8e-05, "loss": 1.682, "step": 4288 }, { "epoch": 0.4635253431319572, "grad_norm": 0.4225302040576935, "learning_rate": 8e-05, "loss": 1.7026, "step": 4289 }, { "epoch": 0.463633416189344, "grad_norm": 0.5033335089683533, "learning_rate": 8e-05, "loss": 1.6909, "step": 4290 }, { "epoch": 0.4637414892467308, "grad_norm": 0.4086191952228546, "learning_rate": 8e-05, "loss": 1.7551, "step": 4291 }, { "epoch": 0.46384956230411756, "grad_norm": 0.40312811732292175, "learning_rate": 8e-05, "loss": 1.7368, "step": 4292 }, { "epoch": 0.46395763536150436, "grad_norm": 0.38233914971351624, "learning_rate": 8e-05, "loss": 1.6473, "step": 4293 }, { "epoch": 0.46406570841889117, "grad_norm": 0.4416559934616089, "learning_rate": 8e-05, "loss": 1.8628, "step": 4294 }, { "epoch": 0.46417378147627797, "grad_norm": 0.39385291934013367, "learning_rate": 8e-05, "loss": 1.7711, "step": 4295 }, { "epoch": 0.4642818545336648, "grad_norm": 0.42280063033103943, "learning_rate": 8e-05, "loss": 1.6657, "step": 4296 }, { "epoch": 0.4643899275910516, "grad_norm": 0.40708550810813904, "learning_rate": 8e-05, "loss": 1.7306, "step": 4297 }, { "epoch": 0.4644980006484383, "grad_norm": 0.40606924891471863, "learning_rate": 8e-05, "loss": 1.7363, "step": 4298 }, { "epoch": 0.46460607370582513, "grad_norm": 0.3782714903354645, "learning_rate": 8e-05, "loss": 1.6565, "step": 4299 }, { "epoch": 0.46471414676321193, "grad_norm": 0.4321131408214569, "learning_rate": 8e-05, "loss": 1.6977, "step": 4300 }, { "epoch": 0.46482221982059874, "grad_norm": 0.42701172828674316, "learning_rate": 8e-05, "loss": 1.7951, "step": 4301 }, { "epoch": 0.46493029287798554, "grad_norm": 0.3887809216976166, "learning_rate": 8e-05, "loss": 1.6972, "step": 4302 }, { "epoch": 0.4650383659353723, "grad_norm": 0.378823459148407, "learning_rate": 8e-05, "loss": 1.7314, "step": 4303 }, { "epoch": 0.4651464389927591, "grad_norm": 0.40994536876678467, "learning_rate": 8e-05, "loss": 1.6515, "step": 4304 }, { "epoch": 0.4652545120501459, "grad_norm": 0.41217079758644104, "learning_rate": 8e-05, "loss": 1.8695, "step": 4305 }, { "epoch": 0.4653625851075327, "grad_norm": 0.4261700510978699, "learning_rate": 8e-05, "loss": 1.8445, "step": 4306 }, { "epoch": 0.4654706581649195, "grad_norm": 0.35033541917800903, "learning_rate": 8e-05, "loss": 1.3345, "step": 4307 }, { "epoch": 0.46557873122230625, "grad_norm": 0.4369705021381378, "learning_rate": 8e-05, "loss": 1.7823, "step": 4308 }, { "epoch": 0.46568680427969306, "grad_norm": 0.396161288022995, "learning_rate": 8e-05, "loss": 1.6919, "step": 4309 }, { "epoch": 0.46579487733707986, "grad_norm": 0.39318424463272095, "learning_rate": 8e-05, "loss": 1.687, "step": 4310 }, { "epoch": 0.46590295039446666, "grad_norm": 0.4040050506591797, "learning_rate": 8e-05, "loss": 1.6674, "step": 4311 }, { "epoch": 0.46601102345185347, "grad_norm": 0.3876594603061676, "learning_rate": 8e-05, "loss": 1.6904, "step": 4312 }, { "epoch": 0.46611909650924027, "grad_norm": 0.40077510476112366, "learning_rate": 8e-05, "loss": 1.736, "step": 4313 }, { "epoch": 0.466227169566627, "grad_norm": 0.3969590961933136, "learning_rate": 8e-05, "loss": 1.639, "step": 4314 }, { "epoch": 0.4663352426240138, "grad_norm": 0.38464590907096863, "learning_rate": 8e-05, "loss": 1.5865, "step": 4315 }, { "epoch": 0.4664433156814006, "grad_norm": 0.3839821517467499, "learning_rate": 8e-05, "loss": 1.7412, "step": 4316 }, { "epoch": 0.46655138873878743, "grad_norm": 0.4156714677810669, "learning_rate": 8e-05, "loss": 1.7886, "step": 4317 }, { "epoch": 0.46665946179617424, "grad_norm": 0.3865569531917572, "learning_rate": 8e-05, "loss": 1.6934, "step": 4318 }, { "epoch": 0.466767534853561, "grad_norm": 0.38035380840301514, "learning_rate": 8e-05, "loss": 1.5075, "step": 4319 }, { "epoch": 0.4668756079109478, "grad_norm": 0.3947802782058716, "learning_rate": 8e-05, "loss": 1.5321, "step": 4320 }, { "epoch": 0.4669836809683346, "grad_norm": 0.3801569938659668, "learning_rate": 8e-05, "loss": 1.7459, "step": 4321 }, { "epoch": 0.4670917540257214, "grad_norm": 0.42639559507369995, "learning_rate": 8e-05, "loss": 1.8251, "step": 4322 }, { "epoch": 0.4671998270831082, "grad_norm": 0.4697808027267456, "learning_rate": 8e-05, "loss": 1.7436, "step": 4323 }, { "epoch": 0.46730790014049495, "grad_norm": 0.39127108454704285, "learning_rate": 8e-05, "loss": 1.5958, "step": 4324 }, { "epoch": 0.46741597319788175, "grad_norm": 0.4393249750137329, "learning_rate": 8e-05, "loss": 1.845, "step": 4325 }, { "epoch": 0.46752404625526855, "grad_norm": 0.3974098265171051, "learning_rate": 8e-05, "loss": 1.7294, "step": 4326 }, { "epoch": 0.46763211931265536, "grad_norm": 0.4098997116088867, "learning_rate": 8e-05, "loss": 1.7576, "step": 4327 }, { "epoch": 0.46774019237004216, "grad_norm": 0.44012629985809326, "learning_rate": 8e-05, "loss": 1.8869, "step": 4328 }, { "epoch": 0.46784826542742897, "grad_norm": 0.43652862310409546, "learning_rate": 8e-05, "loss": 1.7634, "step": 4329 }, { "epoch": 0.4679563384848157, "grad_norm": 0.41287779808044434, "learning_rate": 8e-05, "loss": 1.8242, "step": 4330 }, { "epoch": 0.4680644115422025, "grad_norm": 0.4128740727901459, "learning_rate": 8e-05, "loss": 1.7326, "step": 4331 }, { "epoch": 0.4681724845995893, "grad_norm": 0.42176634073257446, "learning_rate": 8e-05, "loss": 1.8116, "step": 4332 }, { "epoch": 0.4682805576569761, "grad_norm": 0.4075094759464264, "learning_rate": 8e-05, "loss": 1.7959, "step": 4333 }, { "epoch": 0.46838863071436293, "grad_norm": 0.38438481092453003, "learning_rate": 8e-05, "loss": 1.5871, "step": 4334 }, { "epoch": 0.4684967037717497, "grad_norm": 0.3875412046909332, "learning_rate": 8e-05, "loss": 1.6973, "step": 4335 }, { "epoch": 0.4686047768291365, "grad_norm": 0.41900065541267395, "learning_rate": 8e-05, "loss": 1.751, "step": 4336 }, { "epoch": 0.4687128498865233, "grad_norm": 0.38133275508880615, "learning_rate": 8e-05, "loss": 1.5882, "step": 4337 }, { "epoch": 0.4688209229439101, "grad_norm": 0.4026641845703125, "learning_rate": 8e-05, "loss": 1.795, "step": 4338 }, { "epoch": 0.4689289960012969, "grad_norm": 0.40954869985580444, "learning_rate": 8e-05, "loss": 1.8215, "step": 4339 }, { "epoch": 0.4690370690586837, "grad_norm": 0.4353555738925934, "learning_rate": 8e-05, "loss": 1.9223, "step": 4340 }, { "epoch": 0.46914514211607045, "grad_norm": 0.41750815510749817, "learning_rate": 8e-05, "loss": 1.7312, "step": 4341 }, { "epoch": 0.46925321517345725, "grad_norm": 0.5474597215652466, "learning_rate": 8e-05, "loss": 1.9644, "step": 4342 }, { "epoch": 0.46936128823084405, "grad_norm": 0.378177285194397, "learning_rate": 8e-05, "loss": 1.6865, "step": 4343 }, { "epoch": 0.46946936128823086, "grad_norm": 0.4031790792942047, "learning_rate": 8e-05, "loss": 1.7007, "step": 4344 }, { "epoch": 0.46957743434561766, "grad_norm": 0.387382447719574, "learning_rate": 8e-05, "loss": 1.7539, "step": 4345 }, { "epoch": 0.4696855074030044, "grad_norm": 0.41637444496154785, "learning_rate": 8e-05, "loss": 1.6198, "step": 4346 }, { "epoch": 0.4697935804603912, "grad_norm": 0.427708238363266, "learning_rate": 8e-05, "loss": 1.8234, "step": 4347 }, { "epoch": 0.469901653517778, "grad_norm": 0.41436535120010376, "learning_rate": 8e-05, "loss": 1.9051, "step": 4348 }, { "epoch": 0.4700097265751648, "grad_norm": 0.3977288007736206, "learning_rate": 8e-05, "loss": 1.7416, "step": 4349 }, { "epoch": 0.4701177996325516, "grad_norm": 0.3888220191001892, "learning_rate": 8e-05, "loss": 1.6692, "step": 4350 }, { "epoch": 0.4702258726899384, "grad_norm": 0.41765326261520386, "learning_rate": 8e-05, "loss": 1.7674, "step": 4351 }, { "epoch": 0.4703339457473252, "grad_norm": 0.46067631244659424, "learning_rate": 8e-05, "loss": 1.9207, "step": 4352 }, { "epoch": 0.470442018804712, "grad_norm": 0.45710182189941406, "learning_rate": 8e-05, "loss": 1.8075, "step": 4353 }, { "epoch": 0.4705500918620988, "grad_norm": 0.4180136024951935, "learning_rate": 8e-05, "loss": 1.5251, "step": 4354 }, { "epoch": 0.4706581649194856, "grad_norm": 0.398973673582077, "learning_rate": 8e-05, "loss": 1.5873, "step": 4355 }, { "epoch": 0.4707662379768724, "grad_norm": 0.44040340185165405, "learning_rate": 8e-05, "loss": 1.8309, "step": 4356 }, { "epoch": 0.47087431103425914, "grad_norm": 0.39720261096954346, "learning_rate": 8e-05, "loss": 1.8151, "step": 4357 }, { "epoch": 0.47098238409164594, "grad_norm": 0.41815540194511414, "learning_rate": 8e-05, "loss": 1.5681, "step": 4358 }, { "epoch": 0.47109045714903275, "grad_norm": 0.40825673937797546, "learning_rate": 8e-05, "loss": 1.8415, "step": 4359 }, { "epoch": 0.47119853020641955, "grad_norm": 0.3943626582622528, "learning_rate": 8e-05, "loss": 1.664, "step": 4360 }, { "epoch": 0.47130660326380636, "grad_norm": 0.422489732503891, "learning_rate": 8e-05, "loss": 1.9152, "step": 4361 }, { "epoch": 0.4714146763211931, "grad_norm": 0.3940838575363159, "learning_rate": 8e-05, "loss": 1.6895, "step": 4362 }, { "epoch": 0.4715227493785799, "grad_norm": 0.4819503128528595, "learning_rate": 8e-05, "loss": 2.0226, "step": 4363 }, { "epoch": 0.4716308224359667, "grad_norm": 0.433308482170105, "learning_rate": 8e-05, "loss": 1.8545, "step": 4364 }, { "epoch": 0.4717388954933535, "grad_norm": 0.3972829580307007, "learning_rate": 8e-05, "loss": 1.4199, "step": 4365 }, { "epoch": 0.4718469685507403, "grad_norm": 0.4517643451690674, "learning_rate": 8e-05, "loss": 1.6275, "step": 4366 }, { "epoch": 0.47195504160812707, "grad_norm": 0.49456995725631714, "learning_rate": 8e-05, "loss": 2.1072, "step": 4367 }, { "epoch": 0.47206311466551387, "grad_norm": 0.38539162278175354, "learning_rate": 8e-05, "loss": 1.6398, "step": 4368 }, { "epoch": 0.4721711877229007, "grad_norm": 0.4064953029155731, "learning_rate": 8e-05, "loss": 1.7952, "step": 4369 }, { "epoch": 0.4722792607802875, "grad_norm": 0.3970552086830139, "learning_rate": 8e-05, "loss": 1.6494, "step": 4370 }, { "epoch": 0.4723873338376743, "grad_norm": 0.4211111068725586, "learning_rate": 8e-05, "loss": 1.7976, "step": 4371 }, { "epoch": 0.4724954068950611, "grad_norm": 0.3946819603443146, "learning_rate": 8e-05, "loss": 1.6797, "step": 4372 }, { "epoch": 0.47260347995244784, "grad_norm": 0.483755886554718, "learning_rate": 8e-05, "loss": 1.9134, "step": 4373 }, { "epoch": 0.47271155300983464, "grad_norm": 0.3702860474586487, "learning_rate": 8e-05, "loss": 1.6536, "step": 4374 }, { "epoch": 0.47281962606722144, "grad_norm": 0.4280916452407837, "learning_rate": 8e-05, "loss": 1.6624, "step": 4375 }, { "epoch": 0.47292769912460825, "grad_norm": 0.3854413330554962, "learning_rate": 8e-05, "loss": 1.6303, "step": 4376 }, { "epoch": 0.47303577218199505, "grad_norm": 0.41645359992980957, "learning_rate": 8e-05, "loss": 1.77, "step": 4377 }, { "epoch": 0.4731438452393818, "grad_norm": 0.4792899489402771, "learning_rate": 8e-05, "loss": 1.8368, "step": 4378 }, { "epoch": 0.4732519182967686, "grad_norm": 0.40461787581443787, "learning_rate": 8e-05, "loss": 1.769, "step": 4379 }, { "epoch": 0.4733599913541554, "grad_norm": 0.397129625082016, "learning_rate": 8e-05, "loss": 1.6811, "step": 4380 }, { "epoch": 0.4734680644115422, "grad_norm": 0.3850407302379608, "learning_rate": 8e-05, "loss": 1.675, "step": 4381 }, { "epoch": 0.473576137468929, "grad_norm": 0.39792346954345703, "learning_rate": 8e-05, "loss": 1.5907, "step": 4382 }, { "epoch": 0.47368421052631576, "grad_norm": 0.40356749296188354, "learning_rate": 8e-05, "loss": 1.6636, "step": 4383 }, { "epoch": 0.47379228358370257, "grad_norm": 0.39915353059768677, "learning_rate": 8e-05, "loss": 1.8245, "step": 4384 }, { "epoch": 0.47390035664108937, "grad_norm": 0.42239752411842346, "learning_rate": 8e-05, "loss": 1.9393, "step": 4385 }, { "epoch": 0.4740084296984762, "grad_norm": 0.40600547194480896, "learning_rate": 8e-05, "loss": 1.7298, "step": 4386 }, { "epoch": 0.474116502755863, "grad_norm": 0.4127482771873474, "learning_rate": 8e-05, "loss": 1.7381, "step": 4387 }, { "epoch": 0.4742245758132498, "grad_norm": 0.40829917788505554, "learning_rate": 8e-05, "loss": 1.6669, "step": 4388 }, { "epoch": 0.47433264887063653, "grad_norm": 0.4510522782802582, "learning_rate": 8e-05, "loss": 1.9538, "step": 4389 }, { "epoch": 0.47444072192802333, "grad_norm": 0.43200254440307617, "learning_rate": 8e-05, "loss": 1.7665, "step": 4390 }, { "epoch": 0.47454879498541014, "grad_norm": 0.38068196177482605, "learning_rate": 8e-05, "loss": 1.6835, "step": 4391 }, { "epoch": 0.47465686804279694, "grad_norm": 0.38059884309768677, "learning_rate": 8e-05, "loss": 1.7722, "step": 4392 }, { "epoch": 0.47476494110018375, "grad_norm": 0.4309141933917999, "learning_rate": 8e-05, "loss": 1.8183, "step": 4393 }, { "epoch": 0.4748730141575705, "grad_norm": 0.4085610806941986, "learning_rate": 8e-05, "loss": 1.6723, "step": 4394 }, { "epoch": 0.4749810872149573, "grad_norm": 0.3846578896045685, "learning_rate": 8e-05, "loss": 1.6585, "step": 4395 }, { "epoch": 0.4750891602723441, "grad_norm": 0.3819640278816223, "learning_rate": 8e-05, "loss": 1.6943, "step": 4396 }, { "epoch": 0.4751972333297309, "grad_norm": 0.39420387148857117, "learning_rate": 8e-05, "loss": 1.6839, "step": 4397 }, { "epoch": 0.4753053063871177, "grad_norm": 0.40593892335891724, "learning_rate": 8e-05, "loss": 1.7796, "step": 4398 }, { "epoch": 0.47541337944450446, "grad_norm": 0.41556963324546814, "learning_rate": 8e-05, "loss": 1.574, "step": 4399 }, { "epoch": 0.47552145250189126, "grad_norm": 0.41296055912971497, "learning_rate": 8e-05, "loss": 1.778, "step": 4400 }, { "epoch": 0.47562952555927807, "grad_norm": 0.3912244439125061, "learning_rate": 8e-05, "loss": 1.7788, "step": 4401 }, { "epoch": 0.47573759861666487, "grad_norm": 0.3882213532924652, "learning_rate": 8e-05, "loss": 1.7627, "step": 4402 }, { "epoch": 0.4758456716740517, "grad_norm": 0.4012654721736908, "learning_rate": 8e-05, "loss": 1.7493, "step": 4403 }, { "epoch": 0.4759537447314385, "grad_norm": 0.406279057264328, "learning_rate": 8e-05, "loss": 1.8219, "step": 4404 }, { "epoch": 0.4760618177888252, "grad_norm": 0.3669048547744751, "learning_rate": 8e-05, "loss": 1.6331, "step": 4405 }, { "epoch": 0.47616989084621203, "grad_norm": 0.4104841947555542, "learning_rate": 8e-05, "loss": 1.7244, "step": 4406 }, { "epoch": 0.47627796390359883, "grad_norm": 0.4692522883415222, "learning_rate": 8e-05, "loss": 2.0614, "step": 4407 }, { "epoch": 0.47638603696098564, "grad_norm": 0.39436599612236023, "learning_rate": 8e-05, "loss": 1.6795, "step": 4408 }, { "epoch": 0.47649411001837244, "grad_norm": 0.38139328360557556, "learning_rate": 8e-05, "loss": 1.7118, "step": 4409 }, { "epoch": 0.4766021830757592, "grad_norm": 0.3653238117694855, "learning_rate": 8e-05, "loss": 1.7316, "step": 4410 }, { "epoch": 0.476710256133146, "grad_norm": 0.47773492336273193, "learning_rate": 8e-05, "loss": 1.8691, "step": 4411 }, { "epoch": 0.4768183291905328, "grad_norm": 0.41180169582366943, "learning_rate": 8e-05, "loss": 1.5012, "step": 4412 }, { "epoch": 0.4769264022479196, "grad_norm": 0.4035488963127136, "learning_rate": 8e-05, "loss": 1.6536, "step": 4413 }, { "epoch": 0.4770344753053064, "grad_norm": 0.4330061674118042, "learning_rate": 8e-05, "loss": 1.6203, "step": 4414 }, { "epoch": 0.47714254836269315, "grad_norm": 0.4134182035923004, "learning_rate": 8e-05, "loss": 1.7122, "step": 4415 }, { "epoch": 0.47725062142007996, "grad_norm": 0.4109426438808441, "learning_rate": 8e-05, "loss": 1.816, "step": 4416 }, { "epoch": 0.47735869447746676, "grad_norm": 0.3879695534706116, "learning_rate": 8e-05, "loss": 1.5841, "step": 4417 }, { "epoch": 0.47746676753485356, "grad_norm": 0.3969479501247406, "learning_rate": 8e-05, "loss": 1.7279, "step": 4418 }, { "epoch": 0.47757484059224037, "grad_norm": 0.3923637270927429, "learning_rate": 8e-05, "loss": 1.5371, "step": 4419 }, { "epoch": 0.47768291364962717, "grad_norm": 0.37929192185401917, "learning_rate": 8e-05, "loss": 1.6564, "step": 4420 }, { "epoch": 0.4777909867070139, "grad_norm": 0.393903523683548, "learning_rate": 8e-05, "loss": 1.7322, "step": 4421 }, { "epoch": 0.4778990597644007, "grad_norm": 0.3977348506450653, "learning_rate": 8e-05, "loss": 1.5666, "step": 4422 }, { "epoch": 0.47800713282178753, "grad_norm": 0.4265556037425995, "learning_rate": 8e-05, "loss": 1.7529, "step": 4423 }, { "epoch": 0.47811520587917433, "grad_norm": 0.4359314739704132, "learning_rate": 8e-05, "loss": 1.805, "step": 4424 }, { "epoch": 0.47822327893656114, "grad_norm": 0.4290197193622589, "learning_rate": 8e-05, "loss": 1.9265, "step": 4425 }, { "epoch": 0.4783313519939479, "grad_norm": 0.4240277409553528, "learning_rate": 8e-05, "loss": 1.7292, "step": 4426 }, { "epoch": 0.4784394250513347, "grad_norm": 0.42168569564819336, "learning_rate": 8e-05, "loss": 1.8055, "step": 4427 }, { "epoch": 0.4785474981087215, "grad_norm": 0.4279271066188812, "learning_rate": 8e-05, "loss": 1.6262, "step": 4428 }, { "epoch": 0.4786555711661083, "grad_norm": 0.38093265891075134, "learning_rate": 8e-05, "loss": 1.5498, "step": 4429 }, { "epoch": 0.4787636442234951, "grad_norm": 0.43716198205947876, "learning_rate": 8e-05, "loss": 1.8098, "step": 4430 }, { "epoch": 0.4788717172808819, "grad_norm": 0.40469709038734436, "learning_rate": 8e-05, "loss": 1.609, "step": 4431 }, { "epoch": 0.47897979033826865, "grad_norm": 0.4298054277896881, "learning_rate": 8e-05, "loss": 1.7464, "step": 4432 }, { "epoch": 0.47908786339565546, "grad_norm": 0.406124085187912, "learning_rate": 8e-05, "loss": 1.5323, "step": 4433 }, { "epoch": 0.47919593645304226, "grad_norm": 0.39833322167396545, "learning_rate": 8e-05, "loss": 1.7075, "step": 4434 }, { "epoch": 0.47930400951042906, "grad_norm": 0.4070724546909332, "learning_rate": 8e-05, "loss": 1.8449, "step": 4435 }, { "epoch": 0.47941208256781587, "grad_norm": 0.4110445976257324, "learning_rate": 8e-05, "loss": 1.7396, "step": 4436 }, { "epoch": 0.4795201556252026, "grad_norm": 0.44566917419433594, "learning_rate": 8e-05, "loss": 1.7021, "step": 4437 }, { "epoch": 0.4796282286825894, "grad_norm": 0.41855207085609436, "learning_rate": 8e-05, "loss": 1.6947, "step": 4438 }, { "epoch": 0.4797363017399762, "grad_norm": 0.39597100019454956, "learning_rate": 8e-05, "loss": 1.654, "step": 4439 }, { "epoch": 0.479844374797363, "grad_norm": 0.3993816673755646, "learning_rate": 8e-05, "loss": 1.6734, "step": 4440 }, { "epoch": 0.47995244785474983, "grad_norm": 0.37629780173301697, "learning_rate": 8e-05, "loss": 1.5556, "step": 4441 }, { "epoch": 0.4800605209121366, "grad_norm": 0.4099539816379547, "learning_rate": 8e-05, "loss": 1.8828, "step": 4442 }, { "epoch": 0.4801685939695234, "grad_norm": 0.3883780241012573, "learning_rate": 8e-05, "loss": 1.7729, "step": 4443 }, { "epoch": 0.4802766670269102, "grad_norm": 0.3549973964691162, "learning_rate": 8e-05, "loss": 1.6145, "step": 4444 }, { "epoch": 0.480384740084297, "grad_norm": 0.42205509543418884, "learning_rate": 8e-05, "loss": 1.7368, "step": 4445 }, { "epoch": 0.4804928131416838, "grad_norm": 0.4315900206565857, "learning_rate": 8e-05, "loss": 1.8262, "step": 4446 }, { "epoch": 0.4806008861990706, "grad_norm": 0.38961830735206604, "learning_rate": 8e-05, "loss": 1.6552, "step": 4447 }, { "epoch": 0.48070895925645735, "grad_norm": 0.40467625856399536, "learning_rate": 8e-05, "loss": 1.7665, "step": 4448 }, { "epoch": 0.48081703231384415, "grad_norm": 0.4513581693172455, "learning_rate": 8e-05, "loss": 1.7646, "step": 4449 }, { "epoch": 0.48092510537123095, "grad_norm": 0.4321550130844116, "learning_rate": 8e-05, "loss": 1.5715, "step": 4450 }, { "epoch": 0.48103317842861776, "grad_norm": 0.42358481884002686, "learning_rate": 8e-05, "loss": 1.6498, "step": 4451 }, { "epoch": 0.48114125148600456, "grad_norm": 0.461465448141098, "learning_rate": 8e-05, "loss": 1.8365, "step": 4452 }, { "epoch": 0.4812493245433913, "grad_norm": 0.429018497467041, "learning_rate": 8e-05, "loss": 1.8486, "step": 4453 }, { "epoch": 0.4813573976007781, "grad_norm": 0.410231351852417, "learning_rate": 8e-05, "loss": 1.7162, "step": 4454 }, { "epoch": 0.4814654706581649, "grad_norm": 0.4063430428504944, "learning_rate": 8e-05, "loss": 1.6384, "step": 4455 }, { "epoch": 0.4815735437155517, "grad_norm": 0.42961350083351135, "learning_rate": 8e-05, "loss": 1.6931, "step": 4456 }, { "epoch": 0.4816816167729385, "grad_norm": 0.44866907596588135, "learning_rate": 8e-05, "loss": 1.9183, "step": 4457 }, { "epoch": 0.4817896898303253, "grad_norm": 0.42478424310684204, "learning_rate": 8e-05, "loss": 1.7056, "step": 4458 }, { "epoch": 0.4818977628877121, "grad_norm": 0.42392247915267944, "learning_rate": 8e-05, "loss": 1.6186, "step": 4459 }, { "epoch": 0.4820058359450989, "grad_norm": 0.3868848979473114, "learning_rate": 8e-05, "loss": 1.7845, "step": 4460 }, { "epoch": 0.4821139090024857, "grad_norm": 0.426983118057251, "learning_rate": 8e-05, "loss": 1.7918, "step": 4461 }, { "epoch": 0.4822219820598725, "grad_norm": 0.4045998752117157, "learning_rate": 8e-05, "loss": 1.6476, "step": 4462 }, { "epoch": 0.4823300551172593, "grad_norm": 0.37942007184028625, "learning_rate": 8e-05, "loss": 1.7946, "step": 4463 }, { "epoch": 0.48243812817464604, "grad_norm": 0.4016372263431549, "learning_rate": 8e-05, "loss": 1.7546, "step": 4464 }, { "epoch": 0.48254620123203285, "grad_norm": 0.3995134234428406, "learning_rate": 8e-05, "loss": 1.7306, "step": 4465 }, { "epoch": 0.48265427428941965, "grad_norm": 0.3755732476711273, "learning_rate": 8e-05, "loss": 1.6472, "step": 4466 }, { "epoch": 0.48276234734680645, "grad_norm": 0.4377599358558655, "learning_rate": 8e-05, "loss": 1.7974, "step": 4467 }, { "epoch": 0.48287042040419326, "grad_norm": 0.4171310067176819, "learning_rate": 8e-05, "loss": 1.8031, "step": 4468 }, { "epoch": 0.48297849346158, "grad_norm": 0.38294002413749695, "learning_rate": 8e-05, "loss": 1.5684, "step": 4469 }, { "epoch": 0.4830865665189668, "grad_norm": 0.4068509638309479, "learning_rate": 8e-05, "loss": 1.6412, "step": 4470 }, { "epoch": 0.4831946395763536, "grad_norm": 0.3854493200778961, "learning_rate": 8e-05, "loss": 1.5587, "step": 4471 }, { "epoch": 0.4833027126337404, "grad_norm": 0.42035725712776184, "learning_rate": 8e-05, "loss": 1.8951, "step": 4472 }, { "epoch": 0.4834107856911272, "grad_norm": 0.4494016468524933, "learning_rate": 8e-05, "loss": 1.7786, "step": 4473 }, { "epoch": 0.48351885874851397, "grad_norm": 0.38851773738861084, "learning_rate": 8e-05, "loss": 1.6645, "step": 4474 }, { "epoch": 0.4836269318059008, "grad_norm": 0.3658621311187744, "learning_rate": 8e-05, "loss": 1.5165, "step": 4475 }, { "epoch": 0.4837350048632876, "grad_norm": 0.3905554413795471, "learning_rate": 8e-05, "loss": 1.6728, "step": 4476 }, { "epoch": 0.4838430779206744, "grad_norm": 0.4269133508205414, "learning_rate": 8e-05, "loss": 1.611, "step": 4477 }, { "epoch": 0.4839511509780612, "grad_norm": 0.4251437783241272, "learning_rate": 8e-05, "loss": 1.8256, "step": 4478 }, { "epoch": 0.484059224035448, "grad_norm": 0.3766750395298004, "learning_rate": 8e-05, "loss": 1.6897, "step": 4479 }, { "epoch": 0.48416729709283474, "grad_norm": 0.4373146891593933, "learning_rate": 8e-05, "loss": 1.8201, "step": 4480 }, { "epoch": 0.48427537015022154, "grad_norm": 0.4068273901939392, "learning_rate": 8e-05, "loss": 1.6856, "step": 4481 }, { "epoch": 0.48438344320760834, "grad_norm": 0.38094136118888855, "learning_rate": 8e-05, "loss": 1.6455, "step": 4482 }, { "epoch": 0.48449151626499515, "grad_norm": 0.4291408658027649, "learning_rate": 8e-05, "loss": 1.8316, "step": 4483 }, { "epoch": 0.48459958932238195, "grad_norm": 0.38708606362342834, "learning_rate": 8e-05, "loss": 1.5471, "step": 4484 }, { "epoch": 0.4847076623797687, "grad_norm": 0.4210904538631439, "learning_rate": 8e-05, "loss": 1.7247, "step": 4485 }, { "epoch": 0.4848157354371555, "grad_norm": 0.40632113814353943, "learning_rate": 8e-05, "loss": 1.7915, "step": 4486 }, { "epoch": 0.4849238084945423, "grad_norm": 0.41077303886413574, "learning_rate": 8e-05, "loss": 1.725, "step": 4487 }, { "epoch": 0.4850318815519291, "grad_norm": 0.4161292612552643, "learning_rate": 8e-05, "loss": 1.5933, "step": 4488 }, { "epoch": 0.4851399546093159, "grad_norm": 0.39576563239097595, "learning_rate": 8e-05, "loss": 1.6917, "step": 4489 }, { "epoch": 0.48524802766670266, "grad_norm": 0.4352361261844635, "learning_rate": 8e-05, "loss": 1.7289, "step": 4490 }, { "epoch": 0.48535610072408947, "grad_norm": 0.3711824119091034, "learning_rate": 8e-05, "loss": 1.6364, "step": 4491 }, { "epoch": 0.48546417378147627, "grad_norm": 0.4126344621181488, "learning_rate": 8e-05, "loss": 1.7333, "step": 4492 }, { "epoch": 0.4855722468388631, "grad_norm": 0.41216039657592773, "learning_rate": 8e-05, "loss": 1.7926, "step": 4493 }, { "epoch": 0.4856803198962499, "grad_norm": 0.4065435528755188, "learning_rate": 8e-05, "loss": 1.7913, "step": 4494 }, { "epoch": 0.4857883929536367, "grad_norm": 0.3982198238372803, "learning_rate": 8e-05, "loss": 1.6105, "step": 4495 }, { "epoch": 0.48589646601102343, "grad_norm": 0.48088499903678894, "learning_rate": 8e-05, "loss": 2.035, "step": 4496 }, { "epoch": 0.48600453906841024, "grad_norm": 0.4165959656238556, "learning_rate": 8e-05, "loss": 1.6636, "step": 4497 }, { "epoch": 0.48611261212579704, "grad_norm": 0.40088269114494324, "learning_rate": 8e-05, "loss": 1.5413, "step": 4498 }, { "epoch": 0.48622068518318384, "grad_norm": 0.3928964138031006, "learning_rate": 8e-05, "loss": 1.5888, "step": 4499 }, { "epoch": 0.48632875824057065, "grad_norm": 0.3858298361301422, "learning_rate": 8e-05, "loss": 1.6318, "step": 4500 }, { "epoch": 0.4864368312979574, "grad_norm": 0.3981149196624756, "learning_rate": 8e-05, "loss": 1.7318, "step": 4501 }, { "epoch": 0.4865449043553442, "grad_norm": 0.4077165722846985, "learning_rate": 8e-05, "loss": 1.7257, "step": 4502 }, { "epoch": 0.486652977412731, "grad_norm": 0.4125361740589142, "learning_rate": 8e-05, "loss": 1.7651, "step": 4503 }, { "epoch": 0.4867610504701178, "grad_norm": 0.4113052189350128, "learning_rate": 8e-05, "loss": 1.6128, "step": 4504 }, { "epoch": 0.4868691235275046, "grad_norm": 0.4195830225944519, "learning_rate": 8e-05, "loss": 1.8256, "step": 4505 }, { "epoch": 0.4869771965848914, "grad_norm": 0.40131568908691406, "learning_rate": 8e-05, "loss": 1.7544, "step": 4506 }, { "epoch": 0.48708526964227816, "grad_norm": 0.4021151065826416, "learning_rate": 8e-05, "loss": 1.5619, "step": 4507 }, { "epoch": 0.48719334269966497, "grad_norm": 0.3786599934101105, "learning_rate": 8e-05, "loss": 1.6274, "step": 4508 }, { "epoch": 0.48730141575705177, "grad_norm": 0.39171236753463745, "learning_rate": 8e-05, "loss": 1.6944, "step": 4509 }, { "epoch": 0.4874094888144386, "grad_norm": 0.3862670958042145, "learning_rate": 8e-05, "loss": 1.5794, "step": 4510 }, { "epoch": 0.4875175618718254, "grad_norm": 0.3992129862308502, "learning_rate": 8e-05, "loss": 1.6675, "step": 4511 }, { "epoch": 0.4876256349292121, "grad_norm": 0.39308929443359375, "learning_rate": 8e-05, "loss": 1.7396, "step": 4512 }, { "epoch": 0.48773370798659893, "grad_norm": 0.39373594522476196, "learning_rate": 8e-05, "loss": 1.6743, "step": 4513 }, { "epoch": 0.48784178104398573, "grad_norm": 0.39826205372810364, "learning_rate": 8e-05, "loss": 1.6196, "step": 4514 }, { "epoch": 0.48794985410137254, "grad_norm": 0.4778093695640564, "learning_rate": 8e-05, "loss": 1.7328, "step": 4515 }, { "epoch": 0.48805792715875934, "grad_norm": 0.4264270067214966, "learning_rate": 8e-05, "loss": 1.8237, "step": 4516 }, { "epoch": 0.4881660002161461, "grad_norm": 0.39952999353408813, "learning_rate": 8e-05, "loss": 1.6207, "step": 4517 }, { "epoch": 0.4882740732735329, "grad_norm": 0.3912876546382904, "learning_rate": 8e-05, "loss": 1.531, "step": 4518 }, { "epoch": 0.4883821463309197, "grad_norm": 0.4450731873512268, "learning_rate": 8e-05, "loss": 1.8217, "step": 4519 }, { "epoch": 0.4884902193883065, "grad_norm": 0.4129990041255951, "learning_rate": 8e-05, "loss": 1.873, "step": 4520 }, { "epoch": 0.4885982924456933, "grad_norm": 0.38996970653533936, "learning_rate": 8e-05, "loss": 1.7274, "step": 4521 }, { "epoch": 0.4887063655030801, "grad_norm": 0.41124382615089417, "learning_rate": 8e-05, "loss": 1.736, "step": 4522 }, { "epoch": 0.48881443856046686, "grad_norm": 0.4648588001728058, "learning_rate": 8e-05, "loss": 1.8285, "step": 4523 }, { "epoch": 0.48892251161785366, "grad_norm": 0.47971948981285095, "learning_rate": 8e-05, "loss": 1.8545, "step": 4524 }, { "epoch": 0.48903058467524047, "grad_norm": 0.40587490797042847, "learning_rate": 8e-05, "loss": 1.7316, "step": 4525 }, { "epoch": 0.48913865773262727, "grad_norm": 0.43597742915153503, "learning_rate": 8e-05, "loss": 1.7261, "step": 4526 }, { "epoch": 0.4892467307900141, "grad_norm": 0.4638703763484955, "learning_rate": 8e-05, "loss": 1.6373, "step": 4527 }, { "epoch": 0.4893548038474008, "grad_norm": 0.42312783002853394, "learning_rate": 8e-05, "loss": 1.4575, "step": 4528 }, { "epoch": 0.4894628769047876, "grad_norm": 0.39222845435142517, "learning_rate": 8e-05, "loss": 1.7382, "step": 4529 }, { "epoch": 0.48957094996217443, "grad_norm": 0.4099057614803314, "learning_rate": 8e-05, "loss": 1.6386, "step": 4530 }, { "epoch": 0.48967902301956123, "grad_norm": 0.406204491853714, "learning_rate": 8e-05, "loss": 1.6222, "step": 4531 }, { "epoch": 0.48978709607694804, "grad_norm": 0.49371662735939026, "learning_rate": 8e-05, "loss": 1.8274, "step": 4532 }, { "epoch": 0.4898951691343348, "grad_norm": 0.4220612347126007, "learning_rate": 8e-05, "loss": 1.9628, "step": 4533 }, { "epoch": 0.4900032421917216, "grad_norm": 0.3846607506275177, "learning_rate": 8e-05, "loss": 1.6548, "step": 4534 }, { "epoch": 0.4901113152491084, "grad_norm": 0.4358482360839844, "learning_rate": 8e-05, "loss": 1.6209, "step": 4535 }, { "epoch": 0.4902193883064952, "grad_norm": 0.4412643313407898, "learning_rate": 8e-05, "loss": 1.6308, "step": 4536 }, { "epoch": 0.490327461363882, "grad_norm": 0.4089067280292511, "learning_rate": 8e-05, "loss": 1.6897, "step": 4537 }, { "epoch": 0.4904355344212688, "grad_norm": 0.4318627417087555, "learning_rate": 8e-05, "loss": 1.7495, "step": 4538 }, { "epoch": 0.49054360747865555, "grad_norm": 0.41721653938293457, "learning_rate": 8e-05, "loss": 1.5444, "step": 4539 }, { "epoch": 0.49065168053604236, "grad_norm": 0.47165438532829285, "learning_rate": 8e-05, "loss": 1.7536, "step": 4540 }, { "epoch": 0.49075975359342916, "grad_norm": 0.47693413496017456, "learning_rate": 8e-05, "loss": 1.6817, "step": 4541 }, { "epoch": 0.49086782665081596, "grad_norm": 0.3976629972457886, "learning_rate": 8e-05, "loss": 1.7442, "step": 4542 }, { "epoch": 0.49097589970820277, "grad_norm": 0.3838636875152588, "learning_rate": 8e-05, "loss": 1.6901, "step": 4543 }, { "epoch": 0.4910839727655895, "grad_norm": 0.4036515951156616, "learning_rate": 8e-05, "loss": 1.6334, "step": 4544 }, { "epoch": 0.4911920458229763, "grad_norm": 0.4530039429664612, "learning_rate": 8e-05, "loss": 1.8745, "step": 4545 }, { "epoch": 0.4913001188803631, "grad_norm": 0.39521896839141846, "learning_rate": 8e-05, "loss": 1.7036, "step": 4546 }, { "epoch": 0.4914081919377499, "grad_norm": 0.3843677341938019, "learning_rate": 8e-05, "loss": 1.69, "step": 4547 }, { "epoch": 0.49151626499513673, "grad_norm": 0.41856613755226135, "learning_rate": 8e-05, "loss": 1.7805, "step": 4548 }, { "epoch": 0.4916243380525235, "grad_norm": 0.3842417895793915, "learning_rate": 8e-05, "loss": 1.6026, "step": 4549 }, { "epoch": 0.4917324111099103, "grad_norm": 0.41573432087898254, "learning_rate": 8e-05, "loss": 1.7438, "step": 4550 }, { "epoch": 0.4918404841672971, "grad_norm": 0.4763418734073639, "learning_rate": 8e-05, "loss": 1.8534, "step": 4551 }, { "epoch": 0.4919485572246839, "grad_norm": 0.4143596291542053, "learning_rate": 8e-05, "loss": 1.7213, "step": 4552 }, { "epoch": 0.4920566302820707, "grad_norm": 0.43058741092681885, "learning_rate": 8e-05, "loss": 1.7569, "step": 4553 }, { "epoch": 0.4921647033394575, "grad_norm": 0.39957430958747864, "learning_rate": 8e-05, "loss": 1.5292, "step": 4554 }, { "epoch": 0.49227277639684425, "grad_norm": 0.44477221369743347, "learning_rate": 8e-05, "loss": 1.8355, "step": 4555 }, { "epoch": 0.49238084945423105, "grad_norm": 0.42947760224342346, "learning_rate": 8e-05, "loss": 1.6015, "step": 4556 }, { "epoch": 0.49248892251161785, "grad_norm": 0.3672034740447998, "learning_rate": 8e-05, "loss": 1.6425, "step": 4557 }, { "epoch": 0.49259699556900466, "grad_norm": 0.3806307315826416, "learning_rate": 8e-05, "loss": 1.6863, "step": 4558 }, { "epoch": 0.49270506862639146, "grad_norm": 0.3946504294872284, "learning_rate": 8e-05, "loss": 1.4487, "step": 4559 }, { "epoch": 0.4928131416837782, "grad_norm": 0.42138540744781494, "learning_rate": 8e-05, "loss": 1.7622, "step": 4560 }, { "epoch": 0.492921214741165, "grad_norm": 0.4102710485458374, "learning_rate": 8e-05, "loss": 1.7642, "step": 4561 }, { "epoch": 0.4930292877985518, "grad_norm": 0.42016077041625977, "learning_rate": 8e-05, "loss": 1.7404, "step": 4562 }, { "epoch": 0.4931373608559386, "grad_norm": 0.38941633701324463, "learning_rate": 8e-05, "loss": 1.6138, "step": 4563 }, { "epoch": 0.4932454339133254, "grad_norm": 0.4074150621891022, "learning_rate": 8e-05, "loss": 1.6843, "step": 4564 }, { "epoch": 0.4933535069707122, "grad_norm": 0.41428184509277344, "learning_rate": 8e-05, "loss": 1.8069, "step": 4565 }, { "epoch": 0.493461580028099, "grad_norm": 0.4355214536190033, "learning_rate": 8e-05, "loss": 1.8362, "step": 4566 }, { "epoch": 0.4935696530854858, "grad_norm": 0.42752811312675476, "learning_rate": 8e-05, "loss": 1.8528, "step": 4567 }, { "epoch": 0.4936777261428726, "grad_norm": 0.3998211920261383, "learning_rate": 8e-05, "loss": 1.6667, "step": 4568 }, { "epoch": 0.4937857992002594, "grad_norm": 0.42924392223358154, "learning_rate": 8e-05, "loss": 1.8569, "step": 4569 }, { "epoch": 0.4938938722576462, "grad_norm": 0.38225632905960083, "learning_rate": 8e-05, "loss": 1.6995, "step": 4570 }, { "epoch": 0.49400194531503294, "grad_norm": 0.4094962775707245, "learning_rate": 8e-05, "loss": 1.6946, "step": 4571 }, { "epoch": 0.49411001837241975, "grad_norm": 0.38786569237709045, "learning_rate": 8e-05, "loss": 1.7502, "step": 4572 }, { "epoch": 0.49421809142980655, "grad_norm": 0.39534568786621094, "learning_rate": 8e-05, "loss": 1.7299, "step": 4573 }, { "epoch": 0.49432616448719335, "grad_norm": 0.44906845688819885, "learning_rate": 8e-05, "loss": 1.858, "step": 4574 }, { "epoch": 0.49443423754458016, "grad_norm": 0.42588010430336, "learning_rate": 8e-05, "loss": 1.6872, "step": 4575 }, { "epoch": 0.4945423106019669, "grad_norm": 0.39031171798706055, "learning_rate": 8e-05, "loss": 1.6644, "step": 4576 }, { "epoch": 0.4946503836593537, "grad_norm": 0.3910892605781555, "learning_rate": 8e-05, "loss": 1.6419, "step": 4577 }, { "epoch": 0.4947584567167405, "grad_norm": 0.4057159125804901, "learning_rate": 8e-05, "loss": 1.7696, "step": 4578 }, { "epoch": 0.4948665297741273, "grad_norm": 0.3953483998775482, "learning_rate": 8e-05, "loss": 1.6734, "step": 4579 }, { "epoch": 0.4949746028315141, "grad_norm": 0.4400515556335449, "learning_rate": 8e-05, "loss": 1.9922, "step": 4580 }, { "epoch": 0.49508267588890087, "grad_norm": 0.4748576581478119, "learning_rate": 8e-05, "loss": 1.8397, "step": 4581 }, { "epoch": 0.4951907489462877, "grad_norm": 0.4138704538345337, "learning_rate": 8e-05, "loss": 1.7375, "step": 4582 }, { "epoch": 0.4952988220036745, "grad_norm": 0.43630218505859375, "learning_rate": 8e-05, "loss": 1.7078, "step": 4583 }, { "epoch": 0.4954068950610613, "grad_norm": 0.41870805621147156, "learning_rate": 8e-05, "loss": 1.6677, "step": 4584 }, { "epoch": 0.4955149681184481, "grad_norm": 0.4615815579891205, "learning_rate": 8e-05, "loss": 1.8456, "step": 4585 }, { "epoch": 0.4956230411758349, "grad_norm": 0.4198322594165802, "learning_rate": 8e-05, "loss": 1.6839, "step": 4586 }, { "epoch": 0.49573111423322164, "grad_norm": 0.45698797702789307, "learning_rate": 8e-05, "loss": 1.7922, "step": 4587 }, { "epoch": 0.49583918729060844, "grad_norm": 0.426725834608078, "learning_rate": 8e-05, "loss": 1.6433, "step": 4588 }, { "epoch": 0.49594726034799524, "grad_norm": 0.4246196150779724, "learning_rate": 8e-05, "loss": 1.847, "step": 4589 }, { "epoch": 0.49605533340538205, "grad_norm": 0.3808932602405548, "learning_rate": 8e-05, "loss": 1.6011, "step": 4590 }, { "epoch": 0.49616340646276885, "grad_norm": 0.44499140977859497, "learning_rate": 8e-05, "loss": 1.8641, "step": 4591 }, { "epoch": 0.4962714795201556, "grad_norm": 0.4330929219722748, "learning_rate": 8e-05, "loss": 1.6546, "step": 4592 }, { "epoch": 0.4963795525775424, "grad_norm": 0.41983404755592346, "learning_rate": 8e-05, "loss": 1.5631, "step": 4593 }, { "epoch": 0.4964876256349292, "grad_norm": 0.4523801803588867, "learning_rate": 8e-05, "loss": 1.8166, "step": 4594 }, { "epoch": 0.496595698692316, "grad_norm": 0.4522763788700104, "learning_rate": 8e-05, "loss": 1.7319, "step": 4595 }, { "epoch": 0.4967037717497028, "grad_norm": 0.44319260120391846, "learning_rate": 8e-05, "loss": 1.7081, "step": 4596 }, { "epoch": 0.4968118448070896, "grad_norm": 0.4764062166213989, "learning_rate": 8e-05, "loss": 1.8676, "step": 4597 }, { "epoch": 0.49691991786447637, "grad_norm": 0.44301289319992065, "learning_rate": 8e-05, "loss": 1.7514, "step": 4598 }, { "epoch": 0.49702799092186317, "grad_norm": 0.3918417990207672, "learning_rate": 8e-05, "loss": 1.711, "step": 4599 }, { "epoch": 0.49713606397925, "grad_norm": 0.4111935794353485, "learning_rate": 8e-05, "loss": 1.7382, "step": 4600 }, { "epoch": 0.4972441370366368, "grad_norm": 0.40989717841148376, "learning_rate": 8e-05, "loss": 1.629, "step": 4601 }, { "epoch": 0.4973522100940236, "grad_norm": 0.4190944731235504, "learning_rate": 8e-05, "loss": 1.5246, "step": 4602 }, { "epoch": 0.49746028315141033, "grad_norm": 0.43257588148117065, "learning_rate": 8e-05, "loss": 1.6859, "step": 4603 }, { "epoch": 0.49756835620879714, "grad_norm": 0.4143356680870056, "learning_rate": 8e-05, "loss": 1.7318, "step": 4604 }, { "epoch": 0.49767642926618394, "grad_norm": 0.4364134669303894, "learning_rate": 8e-05, "loss": 1.5669, "step": 4605 }, { "epoch": 0.49778450232357074, "grad_norm": 0.47212454676628113, "learning_rate": 8e-05, "loss": 1.8856, "step": 4606 }, { "epoch": 0.49789257538095755, "grad_norm": 0.4265766441822052, "learning_rate": 8e-05, "loss": 1.8067, "step": 4607 }, { "epoch": 0.4980006484383443, "grad_norm": 0.37790244817733765, "learning_rate": 8e-05, "loss": 1.6611, "step": 4608 }, { "epoch": 0.4981087214957311, "grad_norm": 0.4193977117538452, "learning_rate": 8e-05, "loss": 1.7245, "step": 4609 }, { "epoch": 0.4982167945531179, "grad_norm": 0.40634286403656006, "learning_rate": 8e-05, "loss": 1.7468, "step": 4610 }, { "epoch": 0.4983248676105047, "grad_norm": 0.3872389793395996, "learning_rate": 8e-05, "loss": 1.6042, "step": 4611 }, { "epoch": 0.4984329406678915, "grad_norm": 0.43475615978240967, "learning_rate": 8e-05, "loss": 1.652, "step": 4612 }, { "epoch": 0.4985410137252783, "grad_norm": 0.4667968153953552, "learning_rate": 8e-05, "loss": 1.6648, "step": 4613 }, { "epoch": 0.49864908678266506, "grad_norm": 0.3985169231891632, "learning_rate": 8e-05, "loss": 1.6283, "step": 4614 }, { "epoch": 0.49875715984005187, "grad_norm": 0.42451563477516174, "learning_rate": 8e-05, "loss": 1.7535, "step": 4615 }, { "epoch": 0.49886523289743867, "grad_norm": 0.4101034700870514, "learning_rate": 8e-05, "loss": 1.7681, "step": 4616 }, { "epoch": 0.4989733059548255, "grad_norm": 0.3983090817928314, "learning_rate": 8e-05, "loss": 1.907, "step": 4617 }, { "epoch": 0.4990813790122123, "grad_norm": 0.37420305609703064, "learning_rate": 8e-05, "loss": 1.6059, "step": 4618 }, { "epoch": 0.499189452069599, "grad_norm": 0.3963295817375183, "learning_rate": 8e-05, "loss": 1.7792, "step": 4619 }, { "epoch": 0.49929752512698583, "grad_norm": 0.40464502573013306, "learning_rate": 8e-05, "loss": 1.5492, "step": 4620 }, { "epoch": 0.49940559818437263, "grad_norm": 0.40743106603622437, "learning_rate": 8e-05, "loss": 1.7119, "step": 4621 }, { "epoch": 0.49951367124175944, "grad_norm": 0.451715350151062, "learning_rate": 8e-05, "loss": 1.6747, "step": 4622 }, { "epoch": 0.49962174429914624, "grad_norm": 0.40801721811294556, "learning_rate": 8e-05, "loss": 1.6356, "step": 4623 }, { "epoch": 0.499729817356533, "grad_norm": 0.4135707914829254, "learning_rate": 8e-05, "loss": 1.7293, "step": 4624 }, { "epoch": 0.4998378904139198, "grad_norm": 0.4532390236854553, "learning_rate": 8e-05, "loss": 1.7641, "step": 4625 }, { "epoch": 0.4999459634713066, "grad_norm": 0.4245128929615021, "learning_rate": 8e-05, "loss": 1.8619, "step": 4626 }, { "epoch": 0.5000540365286934, "grad_norm": 0.38395196199417114, "learning_rate": 8e-05, "loss": 1.5814, "step": 4627 }, { "epoch": 0.5001621095860802, "grad_norm": 0.4016512632369995, "learning_rate": 8e-05, "loss": 1.6281, "step": 4628 }, { "epoch": 0.500270182643467, "grad_norm": 0.4368300437927246, "learning_rate": 8e-05, "loss": 1.8255, "step": 4629 }, { "epoch": 0.5003782557008538, "grad_norm": 0.4324931800365448, "learning_rate": 8e-05, "loss": 1.7219, "step": 4630 }, { "epoch": 0.5004863287582406, "grad_norm": 0.4427073895931244, "learning_rate": 8e-05, "loss": 1.6986, "step": 4631 }, { "epoch": 0.5005944018156273, "grad_norm": 0.4387059509754181, "learning_rate": 8e-05, "loss": 1.767, "step": 4632 }, { "epoch": 0.5007024748730141, "grad_norm": 0.4788256585597992, "learning_rate": 8e-05, "loss": 1.9927, "step": 4633 }, { "epoch": 0.5008105479304009, "grad_norm": 0.4731398820877075, "learning_rate": 8e-05, "loss": 1.8522, "step": 4634 }, { "epoch": 0.5009186209877877, "grad_norm": 0.3729192614555359, "learning_rate": 8e-05, "loss": 1.627, "step": 4635 }, { "epoch": 0.5010266940451745, "grad_norm": 0.3934325873851776, "learning_rate": 8e-05, "loss": 1.7136, "step": 4636 }, { "epoch": 0.5011347671025613, "grad_norm": 0.5115959048271179, "learning_rate": 8e-05, "loss": 1.8842, "step": 4637 }, { "epoch": 0.5012428401599481, "grad_norm": 0.3994344174861908, "learning_rate": 8e-05, "loss": 1.6874, "step": 4638 }, { "epoch": 0.5013509132173349, "grad_norm": 0.4582684338092804, "learning_rate": 8e-05, "loss": 1.8734, "step": 4639 }, { "epoch": 0.5014589862747217, "grad_norm": 0.40094858407974243, "learning_rate": 8e-05, "loss": 1.8413, "step": 4640 }, { "epoch": 0.5015670593321085, "grad_norm": 0.45019951462745667, "learning_rate": 8e-05, "loss": 1.955, "step": 4641 }, { "epoch": 0.5016751323894953, "grad_norm": 0.4047348201274872, "learning_rate": 8e-05, "loss": 1.6152, "step": 4642 }, { "epoch": 0.501783205446882, "grad_norm": 0.39314234256744385, "learning_rate": 8e-05, "loss": 1.6609, "step": 4643 }, { "epoch": 0.5018912785042688, "grad_norm": 0.4130818247795105, "learning_rate": 8e-05, "loss": 1.5828, "step": 4644 }, { "epoch": 0.5019993515616556, "grad_norm": 0.3810843229293823, "learning_rate": 8e-05, "loss": 1.6552, "step": 4645 }, { "epoch": 0.5021074246190425, "grad_norm": 0.3893740475177765, "learning_rate": 8e-05, "loss": 1.5267, "step": 4646 }, { "epoch": 0.5022154976764293, "grad_norm": 0.38051480054855347, "learning_rate": 8e-05, "loss": 1.5957, "step": 4647 }, { "epoch": 0.5023235707338161, "grad_norm": 0.428196519613266, "learning_rate": 8e-05, "loss": 1.7314, "step": 4648 }, { "epoch": 0.5024316437912029, "grad_norm": 0.4092353582382202, "learning_rate": 8e-05, "loss": 1.828, "step": 4649 }, { "epoch": 0.5025397168485897, "grad_norm": 0.3609974980354309, "learning_rate": 8e-05, "loss": 1.6815, "step": 4650 }, { "epoch": 0.5026477899059765, "grad_norm": 0.4251629412174225, "learning_rate": 8e-05, "loss": 1.8908, "step": 4651 }, { "epoch": 0.5027558629633633, "grad_norm": 0.37392112612724304, "learning_rate": 8e-05, "loss": 1.7282, "step": 4652 }, { "epoch": 0.5028639360207501, "grad_norm": 0.4434810280799866, "learning_rate": 8e-05, "loss": 1.845, "step": 4653 }, { "epoch": 0.5029720090781368, "grad_norm": 0.38593465089797974, "learning_rate": 8e-05, "loss": 1.6134, "step": 4654 }, { "epoch": 0.5030800821355236, "grad_norm": 0.4458652436733246, "learning_rate": 8e-05, "loss": 1.7922, "step": 4655 }, { "epoch": 0.5031881551929104, "grad_norm": 0.4200010299682617, "learning_rate": 8e-05, "loss": 1.7636, "step": 4656 }, { "epoch": 0.5032962282502972, "grad_norm": 0.3839763104915619, "learning_rate": 8e-05, "loss": 1.5582, "step": 4657 }, { "epoch": 0.503404301307684, "grad_norm": 0.4212810695171356, "learning_rate": 8e-05, "loss": 1.5734, "step": 4658 }, { "epoch": 0.5035123743650708, "grad_norm": 0.3777220547199249, "learning_rate": 8e-05, "loss": 1.6457, "step": 4659 }, { "epoch": 0.5036204474224576, "grad_norm": 0.41301804780960083, "learning_rate": 8e-05, "loss": 1.6975, "step": 4660 }, { "epoch": 0.5037285204798444, "grad_norm": 0.4142366349697113, "learning_rate": 8e-05, "loss": 1.7257, "step": 4661 }, { "epoch": 0.5038365935372312, "grad_norm": 0.40339356660842896, "learning_rate": 8e-05, "loss": 1.6962, "step": 4662 }, { "epoch": 0.503944666594618, "grad_norm": 0.4041508436203003, "learning_rate": 8e-05, "loss": 1.6953, "step": 4663 }, { "epoch": 0.5040527396520047, "grad_norm": 0.42368394136428833, "learning_rate": 8e-05, "loss": 1.6276, "step": 4664 }, { "epoch": 0.5041608127093915, "grad_norm": 0.3843143880367279, "learning_rate": 8e-05, "loss": 1.6259, "step": 4665 }, { "epoch": 0.5042688857667783, "grad_norm": 0.4025265872478485, "learning_rate": 8e-05, "loss": 1.6217, "step": 4666 }, { "epoch": 0.5043769588241651, "grad_norm": 0.444685697555542, "learning_rate": 8e-05, "loss": 1.5933, "step": 4667 }, { "epoch": 0.5044850318815519, "grad_norm": 0.43712174892425537, "learning_rate": 8e-05, "loss": 1.8116, "step": 4668 }, { "epoch": 0.5045931049389387, "grad_norm": 0.429677277803421, "learning_rate": 8e-05, "loss": 1.8109, "step": 4669 }, { "epoch": 0.5047011779963255, "grad_norm": 0.3996693193912506, "learning_rate": 8e-05, "loss": 1.6544, "step": 4670 }, { "epoch": 0.5048092510537123, "grad_norm": 0.400955468416214, "learning_rate": 8e-05, "loss": 1.7604, "step": 4671 }, { "epoch": 0.5049173241110991, "grad_norm": 0.39697420597076416, "learning_rate": 8e-05, "loss": 1.6263, "step": 4672 }, { "epoch": 0.5050253971684859, "grad_norm": 0.41106465458869934, "learning_rate": 8e-05, "loss": 1.6128, "step": 4673 }, { "epoch": 0.5051334702258727, "grad_norm": 0.3903358578681946, "learning_rate": 8e-05, "loss": 1.7401, "step": 4674 }, { "epoch": 0.5052415432832594, "grad_norm": 0.4007432162761688, "learning_rate": 8e-05, "loss": 1.8125, "step": 4675 }, { "epoch": 0.5053496163406462, "grad_norm": 0.41746750473976135, "learning_rate": 8e-05, "loss": 1.7796, "step": 4676 }, { "epoch": 0.505457689398033, "grad_norm": 0.4129239320755005, "learning_rate": 8e-05, "loss": 1.7308, "step": 4677 }, { "epoch": 0.5055657624554198, "grad_norm": 0.4130561947822571, "learning_rate": 8e-05, "loss": 1.7485, "step": 4678 }, { "epoch": 0.5056738355128066, "grad_norm": 0.38476482033729553, "learning_rate": 8e-05, "loss": 1.6355, "step": 4679 }, { "epoch": 0.5057819085701935, "grad_norm": 0.41772881150245667, "learning_rate": 8e-05, "loss": 1.7084, "step": 4680 }, { "epoch": 0.5058899816275803, "grad_norm": 0.3783447742462158, "learning_rate": 8e-05, "loss": 1.7684, "step": 4681 }, { "epoch": 0.5059980546849671, "grad_norm": 0.40590929985046387, "learning_rate": 8e-05, "loss": 1.7061, "step": 4682 }, { "epoch": 0.5061061277423539, "grad_norm": 0.41405728459358215, "learning_rate": 8e-05, "loss": 1.7863, "step": 4683 }, { "epoch": 0.5062142007997407, "grad_norm": 0.39419227838516235, "learning_rate": 8e-05, "loss": 1.7967, "step": 4684 }, { "epoch": 0.5063222738571275, "grad_norm": 0.41945794224739075, "learning_rate": 8e-05, "loss": 1.7821, "step": 4685 }, { "epoch": 0.5064303469145142, "grad_norm": 0.43580418825149536, "learning_rate": 8e-05, "loss": 1.7528, "step": 4686 }, { "epoch": 0.506538419971901, "grad_norm": 0.47292235493659973, "learning_rate": 8e-05, "loss": 1.8126, "step": 4687 }, { "epoch": 0.5066464930292878, "grad_norm": 0.3983614444732666, "learning_rate": 8e-05, "loss": 1.5537, "step": 4688 }, { "epoch": 0.5067545660866746, "grad_norm": 0.4345097839832306, "learning_rate": 8e-05, "loss": 1.6542, "step": 4689 }, { "epoch": 0.5068626391440614, "grad_norm": 0.4169449210166931, "learning_rate": 8e-05, "loss": 1.7158, "step": 4690 }, { "epoch": 0.5069707122014482, "grad_norm": 0.3657934367656708, "learning_rate": 8e-05, "loss": 1.5082, "step": 4691 }, { "epoch": 0.507078785258835, "grad_norm": 0.4632113575935364, "learning_rate": 8e-05, "loss": 1.8278, "step": 4692 }, { "epoch": 0.5071868583162218, "grad_norm": 0.3980325162410736, "learning_rate": 8e-05, "loss": 1.7702, "step": 4693 }, { "epoch": 0.5072949313736086, "grad_norm": 0.3889184296131134, "learning_rate": 8e-05, "loss": 1.6634, "step": 4694 }, { "epoch": 0.5074030044309954, "grad_norm": 0.36603844165802, "learning_rate": 8e-05, "loss": 1.4117, "step": 4695 }, { "epoch": 0.5075110774883822, "grad_norm": 0.4158637225627899, "learning_rate": 8e-05, "loss": 1.6991, "step": 4696 }, { "epoch": 0.5076191505457689, "grad_norm": 0.423439621925354, "learning_rate": 8e-05, "loss": 1.7605, "step": 4697 }, { "epoch": 0.5077272236031557, "grad_norm": 0.4030735492706299, "learning_rate": 8e-05, "loss": 1.7774, "step": 4698 }, { "epoch": 0.5078352966605425, "grad_norm": 0.4302954375743866, "learning_rate": 8e-05, "loss": 1.8201, "step": 4699 }, { "epoch": 0.5079433697179293, "grad_norm": 0.3883877098560333, "learning_rate": 8e-05, "loss": 1.6702, "step": 4700 }, { "epoch": 0.5080514427753161, "grad_norm": 0.3819584250450134, "learning_rate": 8e-05, "loss": 1.4745, "step": 4701 }, { "epoch": 0.5081595158327029, "grad_norm": 0.4463454484939575, "learning_rate": 8e-05, "loss": 1.8073, "step": 4702 }, { "epoch": 0.5082675888900897, "grad_norm": 0.39672747254371643, "learning_rate": 8e-05, "loss": 1.6663, "step": 4703 }, { "epoch": 0.5083756619474765, "grad_norm": 0.42274191975593567, "learning_rate": 8e-05, "loss": 1.7293, "step": 4704 }, { "epoch": 0.5084837350048633, "grad_norm": 0.4081788957118988, "learning_rate": 8e-05, "loss": 1.9558, "step": 4705 }, { "epoch": 0.5085918080622501, "grad_norm": 0.4008321762084961, "learning_rate": 8e-05, "loss": 1.7788, "step": 4706 }, { "epoch": 0.5086998811196368, "grad_norm": 0.43986186385154724, "learning_rate": 8e-05, "loss": 1.9007, "step": 4707 }, { "epoch": 0.5088079541770236, "grad_norm": 0.393114298582077, "learning_rate": 8e-05, "loss": 1.6909, "step": 4708 }, { "epoch": 0.5089160272344104, "grad_norm": 0.36888784170150757, "learning_rate": 8e-05, "loss": 1.6337, "step": 4709 }, { "epoch": 0.5090241002917972, "grad_norm": 0.4395067095756531, "learning_rate": 8e-05, "loss": 1.917, "step": 4710 }, { "epoch": 0.509132173349184, "grad_norm": 0.42625027894973755, "learning_rate": 8e-05, "loss": 1.6174, "step": 4711 }, { "epoch": 0.5092402464065708, "grad_norm": 0.41628801822662354, "learning_rate": 8e-05, "loss": 1.5516, "step": 4712 }, { "epoch": 0.5093483194639576, "grad_norm": 0.40050965547561646, "learning_rate": 8e-05, "loss": 1.574, "step": 4713 }, { "epoch": 0.5094563925213444, "grad_norm": 0.4109610915184021, "learning_rate": 8e-05, "loss": 1.7942, "step": 4714 }, { "epoch": 0.5095644655787313, "grad_norm": 0.42685291171073914, "learning_rate": 8e-05, "loss": 1.7486, "step": 4715 }, { "epoch": 0.509672538636118, "grad_norm": 0.39319854974746704, "learning_rate": 8e-05, "loss": 1.605, "step": 4716 }, { "epoch": 0.5097806116935049, "grad_norm": 0.3991563618183136, "learning_rate": 8e-05, "loss": 1.6549, "step": 4717 }, { "epoch": 0.5098886847508916, "grad_norm": 0.39866384863853455, "learning_rate": 8e-05, "loss": 1.6739, "step": 4718 }, { "epoch": 0.5099967578082784, "grad_norm": 0.4868659973144531, "learning_rate": 8e-05, "loss": 1.8198, "step": 4719 }, { "epoch": 0.5101048308656652, "grad_norm": 0.39707428216934204, "learning_rate": 8e-05, "loss": 1.6521, "step": 4720 }, { "epoch": 0.510212903923052, "grad_norm": 0.40801212191581726, "learning_rate": 8e-05, "loss": 1.7363, "step": 4721 }, { "epoch": 0.5103209769804388, "grad_norm": 0.44077005982398987, "learning_rate": 8e-05, "loss": 1.8504, "step": 4722 }, { "epoch": 0.5104290500378256, "grad_norm": 0.3992588520050049, "learning_rate": 8e-05, "loss": 1.7235, "step": 4723 }, { "epoch": 0.5105371230952124, "grad_norm": 0.4055352807044983, "learning_rate": 8e-05, "loss": 1.5592, "step": 4724 }, { "epoch": 0.5106451961525992, "grad_norm": 0.40171900391578674, "learning_rate": 8e-05, "loss": 1.6254, "step": 4725 }, { "epoch": 0.510753269209986, "grad_norm": 0.3749426603317261, "learning_rate": 8e-05, "loss": 1.4668, "step": 4726 }, { "epoch": 0.5108613422673728, "grad_norm": 0.3832828402519226, "learning_rate": 8e-05, "loss": 1.6808, "step": 4727 }, { "epoch": 0.5109694153247596, "grad_norm": 0.4637143611907959, "learning_rate": 8e-05, "loss": 1.9584, "step": 4728 }, { "epoch": 0.5110774883821463, "grad_norm": 0.4723820984363556, "learning_rate": 8e-05, "loss": 1.8783, "step": 4729 }, { "epoch": 0.5111855614395331, "grad_norm": 0.41199734807014465, "learning_rate": 8e-05, "loss": 1.7593, "step": 4730 }, { "epoch": 0.5112936344969199, "grad_norm": 0.38699331879615784, "learning_rate": 8e-05, "loss": 1.5373, "step": 4731 }, { "epoch": 0.5114017075543067, "grad_norm": 0.42380109429359436, "learning_rate": 8e-05, "loss": 1.7024, "step": 4732 }, { "epoch": 0.5115097806116935, "grad_norm": 0.38215172290802, "learning_rate": 8e-05, "loss": 1.7495, "step": 4733 }, { "epoch": 0.5116178536690803, "grad_norm": 0.4330931305885315, "learning_rate": 8e-05, "loss": 1.8796, "step": 4734 }, { "epoch": 0.5117259267264671, "grad_norm": 0.40725767612457275, "learning_rate": 8e-05, "loss": 1.6943, "step": 4735 }, { "epoch": 0.5118339997838539, "grad_norm": 0.4123496413230896, "learning_rate": 8e-05, "loss": 1.6794, "step": 4736 }, { "epoch": 0.5119420728412407, "grad_norm": 0.39777886867523193, "learning_rate": 8e-05, "loss": 1.8141, "step": 4737 }, { "epoch": 0.5120501458986275, "grad_norm": 0.3945496678352356, "learning_rate": 8e-05, "loss": 1.5761, "step": 4738 }, { "epoch": 0.5121582189560142, "grad_norm": 0.42809048295021057, "learning_rate": 8e-05, "loss": 1.7666, "step": 4739 }, { "epoch": 0.512266292013401, "grad_norm": 0.3837582767009735, "learning_rate": 8e-05, "loss": 1.6109, "step": 4740 }, { "epoch": 0.5123743650707878, "grad_norm": 0.4252813458442688, "learning_rate": 8e-05, "loss": 1.9006, "step": 4741 }, { "epoch": 0.5124824381281746, "grad_norm": 0.3779107928276062, "learning_rate": 8e-05, "loss": 1.6411, "step": 4742 }, { "epoch": 0.5125905111855614, "grad_norm": 0.41228488087654114, "learning_rate": 8e-05, "loss": 1.7215, "step": 4743 }, { "epoch": 0.5126985842429482, "grad_norm": 0.40794017910957336, "learning_rate": 8e-05, "loss": 1.7578, "step": 4744 }, { "epoch": 0.512806657300335, "grad_norm": 0.40917718410491943, "learning_rate": 8e-05, "loss": 1.6595, "step": 4745 }, { "epoch": 0.5129147303577218, "grad_norm": 0.3936637341976166, "learning_rate": 8e-05, "loss": 1.6749, "step": 4746 }, { "epoch": 0.5130228034151086, "grad_norm": 0.4197206497192383, "learning_rate": 8e-05, "loss": 1.764, "step": 4747 }, { "epoch": 0.5131308764724954, "grad_norm": 0.44139614701271057, "learning_rate": 8e-05, "loss": 1.7621, "step": 4748 }, { "epoch": 0.5132389495298822, "grad_norm": 0.4017902612686157, "learning_rate": 8e-05, "loss": 1.6353, "step": 4749 }, { "epoch": 0.5133470225872689, "grad_norm": 0.40742015838623047, "learning_rate": 8e-05, "loss": 1.6759, "step": 4750 }, { "epoch": 0.5134550956446557, "grad_norm": 0.43377262353897095, "learning_rate": 8e-05, "loss": 1.7223, "step": 4751 }, { "epoch": 0.5135631687020425, "grad_norm": 0.43455442786216736, "learning_rate": 8e-05, "loss": 1.5976, "step": 4752 }, { "epoch": 0.5136712417594294, "grad_norm": 0.42303210496902466, "learning_rate": 8e-05, "loss": 1.7429, "step": 4753 }, { "epoch": 0.5137793148168162, "grad_norm": 0.4154270589351654, "learning_rate": 8e-05, "loss": 1.8203, "step": 4754 }, { "epoch": 0.513887387874203, "grad_norm": 0.46774837374687195, "learning_rate": 8e-05, "loss": 1.7761, "step": 4755 }, { "epoch": 0.5139954609315898, "grad_norm": 0.40013161301612854, "learning_rate": 8e-05, "loss": 1.8643, "step": 4756 }, { "epoch": 0.5141035339889766, "grad_norm": 0.4226974546909332, "learning_rate": 8e-05, "loss": 1.762, "step": 4757 }, { "epoch": 0.5142116070463634, "grad_norm": 0.629270076751709, "learning_rate": 8e-05, "loss": 1.9884, "step": 4758 }, { "epoch": 0.5143196801037502, "grad_norm": 0.38738080859184265, "learning_rate": 8e-05, "loss": 1.5638, "step": 4759 }, { "epoch": 0.514427753161137, "grad_norm": 0.4483691155910492, "learning_rate": 8e-05, "loss": 1.6217, "step": 4760 }, { "epoch": 0.5145358262185237, "grad_norm": 0.3760409653186798, "learning_rate": 8e-05, "loss": 1.6163, "step": 4761 }, { "epoch": 0.5146438992759105, "grad_norm": 0.5027571320533752, "learning_rate": 8e-05, "loss": 2.0142, "step": 4762 }, { "epoch": 0.5147519723332973, "grad_norm": 0.44235125184059143, "learning_rate": 8e-05, "loss": 1.6542, "step": 4763 }, { "epoch": 0.5148600453906841, "grad_norm": 0.38345348834991455, "learning_rate": 8e-05, "loss": 1.5266, "step": 4764 }, { "epoch": 0.5149681184480709, "grad_norm": 0.41503050923347473, "learning_rate": 8e-05, "loss": 1.7876, "step": 4765 }, { "epoch": 0.5150761915054577, "grad_norm": 0.40613964200019836, "learning_rate": 8e-05, "loss": 1.7009, "step": 4766 }, { "epoch": 0.5151842645628445, "grad_norm": 0.40172886848449707, "learning_rate": 8e-05, "loss": 1.7753, "step": 4767 }, { "epoch": 0.5152923376202313, "grad_norm": 0.43227338790893555, "learning_rate": 8e-05, "loss": 1.7651, "step": 4768 }, { "epoch": 0.5154004106776181, "grad_norm": 0.38273218274116516, "learning_rate": 8e-05, "loss": 1.5844, "step": 4769 }, { "epoch": 0.5155084837350049, "grad_norm": 0.43742454051971436, "learning_rate": 8e-05, "loss": 1.8227, "step": 4770 }, { "epoch": 0.5156165567923917, "grad_norm": 0.3798066973686218, "learning_rate": 8e-05, "loss": 1.6374, "step": 4771 }, { "epoch": 0.5157246298497784, "grad_norm": 0.4344523847103119, "learning_rate": 8e-05, "loss": 1.9303, "step": 4772 }, { "epoch": 0.5158327029071652, "grad_norm": 0.4373953640460968, "learning_rate": 8e-05, "loss": 1.8701, "step": 4773 }, { "epoch": 0.515940775964552, "grad_norm": 0.4018820822238922, "learning_rate": 8e-05, "loss": 1.722, "step": 4774 }, { "epoch": 0.5160488490219388, "grad_norm": 0.4193482995033264, "learning_rate": 8e-05, "loss": 1.7657, "step": 4775 }, { "epoch": 0.5161569220793256, "grad_norm": 0.3853967487812042, "learning_rate": 8e-05, "loss": 1.4658, "step": 4776 }, { "epoch": 0.5162649951367124, "grad_norm": 0.406354695558548, "learning_rate": 8e-05, "loss": 1.5258, "step": 4777 }, { "epoch": 0.5163730681940992, "grad_norm": 0.43898022174835205, "learning_rate": 8e-05, "loss": 1.836, "step": 4778 }, { "epoch": 0.516481141251486, "grad_norm": 0.41223448514938354, "learning_rate": 8e-05, "loss": 1.644, "step": 4779 }, { "epoch": 0.5165892143088728, "grad_norm": 0.43973109126091003, "learning_rate": 8e-05, "loss": 1.8827, "step": 4780 }, { "epoch": 0.5166972873662596, "grad_norm": 0.4527101516723633, "learning_rate": 8e-05, "loss": 1.7622, "step": 4781 }, { "epoch": 0.5168053604236463, "grad_norm": 0.4201156795024872, "learning_rate": 8e-05, "loss": 1.7761, "step": 4782 }, { "epoch": 0.5169134334810331, "grad_norm": 0.4157174825668335, "learning_rate": 8e-05, "loss": 1.6961, "step": 4783 }, { "epoch": 0.5170215065384199, "grad_norm": 0.405518114566803, "learning_rate": 8e-05, "loss": 1.7131, "step": 4784 }, { "epoch": 0.5171295795958067, "grad_norm": 0.4291194975376129, "learning_rate": 8e-05, "loss": 1.8589, "step": 4785 }, { "epoch": 0.5172376526531935, "grad_norm": 0.383748859167099, "learning_rate": 8e-05, "loss": 1.6687, "step": 4786 }, { "epoch": 0.5173457257105804, "grad_norm": 0.3920312821865082, "learning_rate": 8e-05, "loss": 1.7444, "step": 4787 }, { "epoch": 0.5174537987679672, "grad_norm": 0.3934651017189026, "learning_rate": 8e-05, "loss": 1.7991, "step": 4788 }, { "epoch": 0.517561871825354, "grad_norm": 0.38649001717567444, "learning_rate": 8e-05, "loss": 1.6791, "step": 4789 }, { "epoch": 0.5176699448827408, "grad_norm": 0.42291659116744995, "learning_rate": 8e-05, "loss": 1.7796, "step": 4790 }, { "epoch": 0.5177780179401276, "grad_norm": 0.5816870927810669, "learning_rate": 8e-05, "loss": 1.6837, "step": 4791 }, { "epoch": 0.5178860909975144, "grad_norm": 0.4310382902622223, "learning_rate": 8e-05, "loss": 1.8475, "step": 4792 }, { "epoch": 0.5179941640549011, "grad_norm": 0.3854757845401764, "learning_rate": 8e-05, "loss": 1.6007, "step": 4793 }, { "epoch": 0.5181022371122879, "grad_norm": 0.425446480512619, "learning_rate": 8e-05, "loss": 1.867, "step": 4794 }, { "epoch": 0.5182103101696747, "grad_norm": 0.3843328654766083, "learning_rate": 8e-05, "loss": 1.5137, "step": 4795 }, { "epoch": 0.5183183832270615, "grad_norm": 0.4098472595214844, "learning_rate": 8e-05, "loss": 1.6805, "step": 4796 }, { "epoch": 0.5184264562844483, "grad_norm": 0.39224570989608765, "learning_rate": 8e-05, "loss": 1.6642, "step": 4797 }, { "epoch": 0.5185345293418351, "grad_norm": 0.3970671594142914, "learning_rate": 8e-05, "loss": 1.5031, "step": 4798 }, { "epoch": 0.5186426023992219, "grad_norm": 0.3776818513870239, "learning_rate": 8e-05, "loss": 1.4388, "step": 4799 }, { "epoch": 0.5187506754566087, "grad_norm": 0.43284153938293457, "learning_rate": 8e-05, "loss": 1.6793, "step": 4800 }, { "epoch": 0.5188587485139955, "grad_norm": 0.46994057297706604, "learning_rate": 8e-05, "loss": 1.8721, "step": 4801 }, { "epoch": 0.5189668215713823, "grad_norm": 0.43751636147499084, "learning_rate": 8e-05, "loss": 1.8159, "step": 4802 }, { "epoch": 0.5190748946287691, "grad_norm": 0.4508005380630493, "learning_rate": 8e-05, "loss": 1.939, "step": 4803 }, { "epoch": 0.5191829676861558, "grad_norm": 0.4641004502773285, "learning_rate": 8e-05, "loss": 1.9424, "step": 4804 }, { "epoch": 0.5192910407435426, "grad_norm": 0.4278232455253601, "learning_rate": 8e-05, "loss": 1.6807, "step": 4805 }, { "epoch": 0.5193991138009294, "grad_norm": 0.41461819410324097, "learning_rate": 8e-05, "loss": 1.569, "step": 4806 }, { "epoch": 0.5195071868583162, "grad_norm": 0.42840465903282166, "learning_rate": 8e-05, "loss": 1.7524, "step": 4807 }, { "epoch": 0.519615259915703, "grad_norm": 0.4226847290992737, "learning_rate": 8e-05, "loss": 1.6598, "step": 4808 }, { "epoch": 0.5197233329730898, "grad_norm": 0.3884517550468445, "learning_rate": 8e-05, "loss": 1.6587, "step": 4809 }, { "epoch": 0.5198314060304766, "grad_norm": 0.38184890151023865, "learning_rate": 8e-05, "loss": 1.8374, "step": 4810 }, { "epoch": 0.5199394790878634, "grad_norm": 0.40016797184944153, "learning_rate": 8e-05, "loss": 1.7748, "step": 4811 }, { "epoch": 0.5200475521452502, "grad_norm": 0.4112529456615448, "learning_rate": 8e-05, "loss": 1.7414, "step": 4812 }, { "epoch": 0.520155625202637, "grad_norm": 0.3872961401939392, "learning_rate": 8e-05, "loss": 1.7005, "step": 4813 }, { "epoch": 0.5202636982600237, "grad_norm": 0.394014835357666, "learning_rate": 8e-05, "loss": 1.6279, "step": 4814 }, { "epoch": 0.5203717713174105, "grad_norm": 0.3823867440223694, "learning_rate": 8e-05, "loss": 1.5722, "step": 4815 }, { "epoch": 0.5204798443747973, "grad_norm": 0.4088960587978363, "learning_rate": 8e-05, "loss": 1.64, "step": 4816 }, { "epoch": 0.5205879174321841, "grad_norm": 0.44685935974121094, "learning_rate": 8e-05, "loss": 1.8071, "step": 4817 }, { "epoch": 0.5206959904895709, "grad_norm": 0.4237087666988373, "learning_rate": 8e-05, "loss": 1.8672, "step": 4818 }, { "epoch": 0.5208040635469577, "grad_norm": 0.3990110456943512, "learning_rate": 8e-05, "loss": 1.8518, "step": 4819 }, { "epoch": 0.5209121366043445, "grad_norm": 0.41446077823638916, "learning_rate": 8e-05, "loss": 1.6276, "step": 4820 }, { "epoch": 0.5210202096617313, "grad_norm": 0.42045265436172485, "learning_rate": 8e-05, "loss": 1.9048, "step": 4821 }, { "epoch": 0.5211282827191182, "grad_norm": 0.4046669602394104, "learning_rate": 8e-05, "loss": 1.639, "step": 4822 }, { "epoch": 0.521236355776505, "grad_norm": 0.3751082122325897, "learning_rate": 8e-05, "loss": 1.6782, "step": 4823 }, { "epoch": 0.5213444288338918, "grad_norm": 0.3985498547554016, "learning_rate": 8e-05, "loss": 1.6451, "step": 4824 }, { "epoch": 0.5214525018912785, "grad_norm": 0.38252684473991394, "learning_rate": 8e-05, "loss": 1.5875, "step": 4825 }, { "epoch": 0.5215605749486653, "grad_norm": 0.4107401669025421, "learning_rate": 8e-05, "loss": 1.5613, "step": 4826 }, { "epoch": 0.5216686480060521, "grad_norm": 0.45246565341949463, "learning_rate": 8e-05, "loss": 1.771, "step": 4827 }, { "epoch": 0.5217767210634389, "grad_norm": 0.38584524393081665, "learning_rate": 8e-05, "loss": 1.6878, "step": 4828 }, { "epoch": 0.5218847941208257, "grad_norm": 0.42803874611854553, "learning_rate": 8e-05, "loss": 1.8083, "step": 4829 }, { "epoch": 0.5219928671782125, "grad_norm": 0.4374076724052429, "learning_rate": 8e-05, "loss": 1.7597, "step": 4830 }, { "epoch": 0.5221009402355993, "grad_norm": 0.48844432830810547, "learning_rate": 8e-05, "loss": 1.9372, "step": 4831 }, { "epoch": 0.5222090132929861, "grad_norm": 0.42745932936668396, "learning_rate": 8e-05, "loss": 1.6969, "step": 4832 }, { "epoch": 0.5223170863503729, "grad_norm": 0.41262972354888916, "learning_rate": 8e-05, "loss": 1.7552, "step": 4833 }, { "epoch": 0.5224251594077597, "grad_norm": 0.40279683470726013, "learning_rate": 8e-05, "loss": 1.7008, "step": 4834 }, { "epoch": 0.5225332324651465, "grad_norm": 0.42590004205703735, "learning_rate": 8e-05, "loss": 1.7784, "step": 4835 }, { "epoch": 0.5226413055225332, "grad_norm": 0.42654383182525635, "learning_rate": 8e-05, "loss": 1.7202, "step": 4836 }, { "epoch": 0.52274937857992, "grad_norm": 0.4391385316848755, "learning_rate": 8e-05, "loss": 1.7712, "step": 4837 }, { "epoch": 0.5228574516373068, "grad_norm": 0.4117557406425476, "learning_rate": 8e-05, "loss": 1.7483, "step": 4838 }, { "epoch": 0.5229655246946936, "grad_norm": 0.4126327633857727, "learning_rate": 8e-05, "loss": 1.6878, "step": 4839 }, { "epoch": 0.5230735977520804, "grad_norm": 0.46406158804893494, "learning_rate": 8e-05, "loss": 1.8207, "step": 4840 }, { "epoch": 0.5231816708094672, "grad_norm": 0.4451543688774109, "learning_rate": 8e-05, "loss": 1.9556, "step": 4841 }, { "epoch": 0.523289743866854, "grad_norm": 0.39060765504837036, "learning_rate": 8e-05, "loss": 1.6436, "step": 4842 }, { "epoch": 0.5233978169242408, "grad_norm": 0.3856661915779114, "learning_rate": 8e-05, "loss": 1.6003, "step": 4843 }, { "epoch": 0.5235058899816276, "grad_norm": 0.41874486207962036, "learning_rate": 8e-05, "loss": 1.7198, "step": 4844 }, { "epoch": 0.5236139630390144, "grad_norm": 0.38177576661109924, "learning_rate": 8e-05, "loss": 1.7114, "step": 4845 }, { "epoch": 0.5237220360964012, "grad_norm": 0.44675570726394653, "learning_rate": 8e-05, "loss": 1.6544, "step": 4846 }, { "epoch": 0.5238301091537879, "grad_norm": 0.39139798283576965, "learning_rate": 8e-05, "loss": 1.5859, "step": 4847 }, { "epoch": 0.5239381822111747, "grad_norm": 0.37541764974594116, "learning_rate": 8e-05, "loss": 1.5277, "step": 4848 }, { "epoch": 0.5240462552685615, "grad_norm": 0.4628608822822571, "learning_rate": 8e-05, "loss": 1.8211, "step": 4849 }, { "epoch": 0.5241543283259483, "grad_norm": 0.43806251883506775, "learning_rate": 8e-05, "loss": 1.7695, "step": 4850 }, { "epoch": 0.5242624013833351, "grad_norm": 0.45326513051986694, "learning_rate": 8e-05, "loss": 1.61, "step": 4851 }, { "epoch": 0.5243704744407219, "grad_norm": 0.392219603061676, "learning_rate": 8e-05, "loss": 1.7151, "step": 4852 }, { "epoch": 0.5244785474981087, "grad_norm": 0.4122357964515686, "learning_rate": 8e-05, "loss": 1.7077, "step": 4853 }, { "epoch": 0.5245866205554955, "grad_norm": 0.43415403366088867, "learning_rate": 8e-05, "loss": 1.8575, "step": 4854 }, { "epoch": 0.5246946936128823, "grad_norm": 0.3981955349445343, "learning_rate": 8e-05, "loss": 1.6174, "step": 4855 }, { "epoch": 0.5248027666702691, "grad_norm": 0.3884790241718292, "learning_rate": 8e-05, "loss": 1.649, "step": 4856 }, { "epoch": 0.5249108397276558, "grad_norm": 0.39559316635131836, "learning_rate": 8e-05, "loss": 1.6578, "step": 4857 }, { "epoch": 0.5250189127850426, "grad_norm": 0.4003053307533264, "learning_rate": 8e-05, "loss": 1.6516, "step": 4858 }, { "epoch": 0.5251269858424295, "grad_norm": 0.37234529852867126, "learning_rate": 8e-05, "loss": 1.637, "step": 4859 }, { "epoch": 0.5252350588998163, "grad_norm": 0.4073370099067688, "learning_rate": 8e-05, "loss": 1.7174, "step": 4860 }, { "epoch": 0.5253431319572031, "grad_norm": 0.41446778178215027, "learning_rate": 8e-05, "loss": 1.794, "step": 4861 }, { "epoch": 0.5254512050145899, "grad_norm": 0.39679303765296936, "learning_rate": 8e-05, "loss": 1.5728, "step": 4862 }, { "epoch": 0.5255592780719767, "grad_norm": 0.45979782938957214, "learning_rate": 8e-05, "loss": 1.7414, "step": 4863 }, { "epoch": 0.5256673511293635, "grad_norm": 0.4084329605102539, "learning_rate": 8e-05, "loss": 1.7752, "step": 4864 }, { "epoch": 0.5257754241867503, "grad_norm": 0.39365530014038086, "learning_rate": 8e-05, "loss": 1.7289, "step": 4865 }, { "epoch": 0.5258834972441371, "grad_norm": 0.39535269141197205, "learning_rate": 8e-05, "loss": 1.5131, "step": 4866 }, { "epoch": 0.5259915703015239, "grad_norm": 0.4394475817680359, "learning_rate": 8e-05, "loss": 1.8352, "step": 4867 }, { "epoch": 0.5260996433589106, "grad_norm": 0.437294602394104, "learning_rate": 8e-05, "loss": 1.8267, "step": 4868 }, { "epoch": 0.5262077164162974, "grad_norm": 0.4125720262527466, "learning_rate": 8e-05, "loss": 1.6659, "step": 4869 }, { "epoch": 0.5263157894736842, "grad_norm": 0.3946208357810974, "learning_rate": 8e-05, "loss": 1.6276, "step": 4870 }, { "epoch": 0.526423862531071, "grad_norm": 0.4423826336860657, "learning_rate": 8e-05, "loss": 1.6709, "step": 4871 }, { "epoch": 0.5265319355884578, "grad_norm": 0.4441356360912323, "learning_rate": 8e-05, "loss": 1.8176, "step": 4872 }, { "epoch": 0.5266400086458446, "grad_norm": 0.4120944142341614, "learning_rate": 8e-05, "loss": 1.7228, "step": 4873 }, { "epoch": 0.5267480817032314, "grad_norm": 0.43329861760139465, "learning_rate": 8e-05, "loss": 1.82, "step": 4874 }, { "epoch": 0.5268561547606182, "grad_norm": 0.44920217990875244, "learning_rate": 8e-05, "loss": 1.7678, "step": 4875 }, { "epoch": 0.526964227818005, "grad_norm": 0.38295778632164, "learning_rate": 8e-05, "loss": 1.5966, "step": 4876 }, { "epoch": 0.5270723008753918, "grad_norm": 0.40596696734428406, "learning_rate": 8e-05, "loss": 1.7301, "step": 4877 }, { "epoch": 0.5271803739327786, "grad_norm": 0.43769344687461853, "learning_rate": 8e-05, "loss": 1.9155, "step": 4878 }, { "epoch": 0.5272884469901653, "grad_norm": 0.443312406539917, "learning_rate": 8e-05, "loss": 1.9554, "step": 4879 }, { "epoch": 0.5273965200475521, "grad_norm": 0.40869608521461487, "learning_rate": 8e-05, "loss": 1.6775, "step": 4880 }, { "epoch": 0.5275045931049389, "grad_norm": 0.4429347515106201, "learning_rate": 8e-05, "loss": 1.6865, "step": 4881 }, { "epoch": 0.5276126661623257, "grad_norm": 0.42774441838264465, "learning_rate": 8e-05, "loss": 1.8157, "step": 4882 }, { "epoch": 0.5277207392197125, "grad_norm": 0.40076056122779846, "learning_rate": 8e-05, "loss": 1.5737, "step": 4883 }, { "epoch": 0.5278288122770993, "grad_norm": 0.39535146951675415, "learning_rate": 8e-05, "loss": 1.642, "step": 4884 }, { "epoch": 0.5279368853344861, "grad_norm": 0.42379406094551086, "learning_rate": 8e-05, "loss": 1.5001, "step": 4885 }, { "epoch": 0.5280449583918729, "grad_norm": 0.4041021466255188, "learning_rate": 8e-05, "loss": 1.7957, "step": 4886 }, { "epoch": 0.5281530314492597, "grad_norm": 0.4110598564147949, "learning_rate": 8e-05, "loss": 1.6738, "step": 4887 }, { "epoch": 0.5282611045066465, "grad_norm": 0.40826529264450073, "learning_rate": 8e-05, "loss": 1.6994, "step": 4888 }, { "epoch": 0.5283691775640332, "grad_norm": 0.43995004892349243, "learning_rate": 8e-05, "loss": 1.8347, "step": 4889 }, { "epoch": 0.52847725062142, "grad_norm": 0.4595677852630615, "learning_rate": 8e-05, "loss": 1.7604, "step": 4890 }, { "epoch": 0.5285853236788068, "grad_norm": 0.37943774461746216, "learning_rate": 8e-05, "loss": 1.5925, "step": 4891 }, { "epoch": 0.5286933967361936, "grad_norm": 0.45692679286003113, "learning_rate": 8e-05, "loss": 1.6791, "step": 4892 }, { "epoch": 0.5288014697935804, "grad_norm": 0.4138285219669342, "learning_rate": 8e-05, "loss": 1.5395, "step": 4893 }, { "epoch": 0.5289095428509673, "grad_norm": 0.39168405532836914, "learning_rate": 8e-05, "loss": 1.6356, "step": 4894 }, { "epoch": 0.529017615908354, "grad_norm": 0.43939557671546936, "learning_rate": 8e-05, "loss": 1.8134, "step": 4895 }, { "epoch": 0.5291256889657409, "grad_norm": 0.45130473375320435, "learning_rate": 8e-05, "loss": 1.7855, "step": 4896 }, { "epoch": 0.5292337620231277, "grad_norm": 0.3797385096549988, "learning_rate": 8e-05, "loss": 1.6711, "step": 4897 }, { "epoch": 0.5293418350805145, "grad_norm": 0.406331330537796, "learning_rate": 8e-05, "loss": 1.6741, "step": 4898 }, { "epoch": 0.5294499081379013, "grad_norm": 0.3852972090244293, "learning_rate": 8e-05, "loss": 1.7018, "step": 4899 }, { "epoch": 0.529557981195288, "grad_norm": 0.409662663936615, "learning_rate": 8e-05, "loss": 1.7175, "step": 4900 }, { "epoch": 0.5296660542526748, "grad_norm": 0.46197110414505005, "learning_rate": 8e-05, "loss": 1.879, "step": 4901 }, { "epoch": 0.5297741273100616, "grad_norm": 0.39654678106307983, "learning_rate": 8e-05, "loss": 1.8084, "step": 4902 }, { "epoch": 0.5298822003674484, "grad_norm": 0.4031100273132324, "learning_rate": 8e-05, "loss": 1.5731, "step": 4903 }, { "epoch": 0.5299902734248352, "grad_norm": 0.4263877272605896, "learning_rate": 8e-05, "loss": 1.7458, "step": 4904 }, { "epoch": 0.530098346482222, "grad_norm": 0.49063023924827576, "learning_rate": 8e-05, "loss": 1.8162, "step": 4905 }, { "epoch": 0.5302064195396088, "grad_norm": 0.4183445870876312, "learning_rate": 8e-05, "loss": 1.6695, "step": 4906 }, { "epoch": 0.5303144925969956, "grad_norm": 0.3978036344051361, "learning_rate": 8e-05, "loss": 1.7124, "step": 4907 }, { "epoch": 0.5304225656543824, "grad_norm": 0.43422481417655945, "learning_rate": 8e-05, "loss": 1.7773, "step": 4908 }, { "epoch": 0.5305306387117692, "grad_norm": 0.401109516620636, "learning_rate": 8e-05, "loss": 1.7274, "step": 4909 }, { "epoch": 0.530638711769156, "grad_norm": 0.40551817417144775, "learning_rate": 8e-05, "loss": 1.6446, "step": 4910 }, { "epoch": 0.5307467848265427, "grad_norm": 0.40497076511383057, "learning_rate": 8e-05, "loss": 1.7423, "step": 4911 }, { "epoch": 0.5308548578839295, "grad_norm": 0.39855679869651794, "learning_rate": 8e-05, "loss": 1.5437, "step": 4912 }, { "epoch": 0.5309629309413163, "grad_norm": 0.42813795804977417, "learning_rate": 8e-05, "loss": 1.793, "step": 4913 }, { "epoch": 0.5310710039987031, "grad_norm": 0.4858434200286865, "learning_rate": 8e-05, "loss": 1.6226, "step": 4914 }, { "epoch": 0.5311790770560899, "grad_norm": 0.3987038731575012, "learning_rate": 8e-05, "loss": 1.4772, "step": 4915 }, { "epoch": 0.5312871501134767, "grad_norm": 0.4694141745567322, "learning_rate": 8e-05, "loss": 1.9502, "step": 4916 }, { "epoch": 0.5313952231708635, "grad_norm": 0.45072418451309204, "learning_rate": 8e-05, "loss": 1.7758, "step": 4917 }, { "epoch": 0.5315032962282503, "grad_norm": 0.42374247312545776, "learning_rate": 8e-05, "loss": 1.8641, "step": 4918 }, { "epoch": 0.5316113692856371, "grad_norm": 0.4133841395378113, "learning_rate": 8e-05, "loss": 1.8016, "step": 4919 }, { "epoch": 0.5317194423430239, "grad_norm": 0.40989622473716736, "learning_rate": 8e-05, "loss": 1.646, "step": 4920 }, { "epoch": 0.5318275154004107, "grad_norm": 0.43784958124160767, "learning_rate": 8e-05, "loss": 1.7282, "step": 4921 }, { "epoch": 0.5319355884577974, "grad_norm": 0.3818674683570862, "learning_rate": 8e-05, "loss": 1.5484, "step": 4922 }, { "epoch": 0.5320436615151842, "grad_norm": 0.3546164035797119, "learning_rate": 8e-05, "loss": 1.54, "step": 4923 }, { "epoch": 0.532151734572571, "grad_norm": 0.36899980902671814, "learning_rate": 8e-05, "loss": 1.6296, "step": 4924 }, { "epoch": 0.5322598076299578, "grad_norm": 0.4447019696235657, "learning_rate": 8e-05, "loss": 1.8838, "step": 4925 }, { "epoch": 0.5323678806873446, "grad_norm": 0.4012913405895233, "learning_rate": 8e-05, "loss": 1.6091, "step": 4926 }, { "epoch": 0.5324759537447314, "grad_norm": 0.4034128189086914, "learning_rate": 8e-05, "loss": 1.7085, "step": 4927 }, { "epoch": 0.5325840268021182, "grad_norm": 0.3941476345062256, "learning_rate": 8e-05, "loss": 1.7647, "step": 4928 }, { "epoch": 0.532692099859505, "grad_norm": 0.5080128908157349, "learning_rate": 8e-05, "loss": 1.8217, "step": 4929 }, { "epoch": 0.5328001729168919, "grad_norm": 0.41325342655181885, "learning_rate": 8e-05, "loss": 1.8287, "step": 4930 }, { "epoch": 0.5329082459742787, "grad_norm": 0.410104364156723, "learning_rate": 8e-05, "loss": 1.7823, "step": 4931 }, { "epoch": 0.5330163190316654, "grad_norm": 0.4191218316555023, "learning_rate": 8e-05, "loss": 1.708, "step": 4932 }, { "epoch": 0.5331243920890522, "grad_norm": 0.41406193375587463, "learning_rate": 8e-05, "loss": 1.733, "step": 4933 }, { "epoch": 0.533232465146439, "grad_norm": 0.3879591226577759, "learning_rate": 8e-05, "loss": 1.5891, "step": 4934 }, { "epoch": 0.5333405382038258, "grad_norm": 0.44343945384025574, "learning_rate": 8e-05, "loss": 1.73, "step": 4935 }, { "epoch": 0.5334486112612126, "grad_norm": 0.4227713644504547, "learning_rate": 8e-05, "loss": 1.8518, "step": 4936 }, { "epoch": 0.5335566843185994, "grad_norm": 0.38836869597435, "learning_rate": 8e-05, "loss": 1.6848, "step": 4937 }, { "epoch": 0.5336647573759862, "grad_norm": 0.4058745205402374, "learning_rate": 8e-05, "loss": 1.6125, "step": 4938 }, { "epoch": 0.533772830433373, "grad_norm": 0.4631666839122772, "learning_rate": 8e-05, "loss": 1.6722, "step": 4939 }, { "epoch": 0.5338809034907598, "grad_norm": 0.4189152717590332, "learning_rate": 8e-05, "loss": 1.7878, "step": 4940 }, { "epoch": 0.5339889765481466, "grad_norm": 0.42341500520706177, "learning_rate": 8e-05, "loss": 1.8066, "step": 4941 }, { "epoch": 0.5340970496055334, "grad_norm": 0.4770244061946869, "learning_rate": 8e-05, "loss": 1.8376, "step": 4942 }, { "epoch": 0.5342051226629201, "grad_norm": 0.42806220054626465, "learning_rate": 8e-05, "loss": 1.5796, "step": 4943 }, { "epoch": 0.5343131957203069, "grad_norm": 0.40327227115631104, "learning_rate": 8e-05, "loss": 1.5097, "step": 4944 }, { "epoch": 0.5344212687776937, "grad_norm": 0.376825749874115, "learning_rate": 8e-05, "loss": 1.6639, "step": 4945 }, { "epoch": 0.5345293418350805, "grad_norm": 0.4043944478034973, "learning_rate": 8e-05, "loss": 1.5462, "step": 4946 }, { "epoch": 0.5346374148924673, "grad_norm": 0.4299601912498474, "learning_rate": 8e-05, "loss": 1.8006, "step": 4947 }, { "epoch": 0.5347454879498541, "grad_norm": 0.3913020193576813, "learning_rate": 8e-05, "loss": 1.6188, "step": 4948 }, { "epoch": 0.5348535610072409, "grad_norm": 0.42924466729164124, "learning_rate": 8e-05, "loss": 1.8771, "step": 4949 }, { "epoch": 0.5349616340646277, "grad_norm": 0.41477829217910767, "learning_rate": 8e-05, "loss": 1.6529, "step": 4950 }, { "epoch": 0.5350697071220145, "grad_norm": 0.442656010389328, "learning_rate": 8e-05, "loss": 1.6115, "step": 4951 }, { "epoch": 0.5351777801794013, "grad_norm": 0.39936596155166626, "learning_rate": 8e-05, "loss": 1.6037, "step": 4952 }, { "epoch": 0.5352858532367881, "grad_norm": 0.4142223596572876, "learning_rate": 8e-05, "loss": 1.6794, "step": 4953 }, { "epoch": 0.5353939262941748, "grad_norm": 0.38259631395339966, "learning_rate": 8e-05, "loss": 1.6751, "step": 4954 }, { "epoch": 0.5355019993515616, "grad_norm": 0.4126985967159271, "learning_rate": 8e-05, "loss": 1.7393, "step": 4955 }, { "epoch": 0.5356100724089484, "grad_norm": 0.4238603711128235, "learning_rate": 8e-05, "loss": 1.7354, "step": 4956 }, { "epoch": 0.5357181454663352, "grad_norm": 0.4083351790904999, "learning_rate": 8e-05, "loss": 1.5988, "step": 4957 }, { "epoch": 0.535826218523722, "grad_norm": 0.40220242738723755, "learning_rate": 8e-05, "loss": 1.705, "step": 4958 }, { "epoch": 0.5359342915811088, "grad_norm": 0.40210792422294617, "learning_rate": 8e-05, "loss": 1.7388, "step": 4959 }, { "epoch": 0.5360423646384956, "grad_norm": 0.4350391924381256, "learning_rate": 8e-05, "loss": 1.7287, "step": 4960 }, { "epoch": 0.5361504376958824, "grad_norm": 0.4623953402042389, "learning_rate": 8e-05, "loss": 1.8561, "step": 4961 }, { "epoch": 0.5362585107532692, "grad_norm": 0.39749476313591003, "learning_rate": 8e-05, "loss": 1.674, "step": 4962 }, { "epoch": 0.536366583810656, "grad_norm": 0.4659527838230133, "learning_rate": 8e-05, "loss": 1.699, "step": 4963 }, { "epoch": 0.5364746568680427, "grad_norm": 0.46200257539749146, "learning_rate": 8e-05, "loss": 1.7666, "step": 4964 }, { "epoch": 0.5365827299254295, "grad_norm": 0.4098494052886963, "learning_rate": 8e-05, "loss": 1.7086, "step": 4965 }, { "epoch": 0.5366908029828164, "grad_norm": 0.41835978627204895, "learning_rate": 8e-05, "loss": 1.7408, "step": 4966 }, { "epoch": 0.5367988760402032, "grad_norm": 0.39188632369041443, "learning_rate": 8e-05, "loss": 1.8317, "step": 4967 }, { "epoch": 0.53690694909759, "grad_norm": 0.39533039927482605, "learning_rate": 8e-05, "loss": 1.7549, "step": 4968 }, { "epoch": 0.5370150221549768, "grad_norm": 0.42061132192611694, "learning_rate": 8e-05, "loss": 1.723, "step": 4969 }, { "epoch": 0.5371230952123636, "grad_norm": 0.43010419607162476, "learning_rate": 8e-05, "loss": 1.8395, "step": 4970 }, { "epoch": 0.5372311682697504, "grad_norm": 0.38996654748916626, "learning_rate": 8e-05, "loss": 1.6134, "step": 4971 }, { "epoch": 0.5373392413271372, "grad_norm": 0.41649290919303894, "learning_rate": 8e-05, "loss": 1.6993, "step": 4972 }, { "epoch": 0.537447314384524, "grad_norm": 0.4055330157279968, "learning_rate": 8e-05, "loss": 1.6144, "step": 4973 }, { "epoch": 0.5375553874419108, "grad_norm": 0.408184677362442, "learning_rate": 8e-05, "loss": 1.7796, "step": 4974 }, { "epoch": 0.5376634604992975, "grad_norm": 0.3943346440792084, "learning_rate": 8e-05, "loss": 1.6991, "step": 4975 }, { "epoch": 0.5377715335566843, "grad_norm": 0.4003731608390808, "learning_rate": 8e-05, "loss": 1.7364, "step": 4976 }, { "epoch": 0.5378796066140711, "grad_norm": 0.41543933749198914, "learning_rate": 8e-05, "loss": 1.5941, "step": 4977 }, { "epoch": 0.5379876796714579, "grad_norm": 0.4118979871273041, "learning_rate": 8e-05, "loss": 1.7209, "step": 4978 }, { "epoch": 0.5380957527288447, "grad_norm": 0.4620661735534668, "learning_rate": 8e-05, "loss": 1.7529, "step": 4979 }, { "epoch": 0.5382038257862315, "grad_norm": 0.41683709621429443, "learning_rate": 8e-05, "loss": 1.7604, "step": 4980 }, { "epoch": 0.5383118988436183, "grad_norm": 0.37552160024642944, "learning_rate": 8e-05, "loss": 1.5967, "step": 4981 }, { "epoch": 0.5384199719010051, "grad_norm": 0.40685006976127625, "learning_rate": 8e-05, "loss": 1.5256, "step": 4982 }, { "epoch": 0.5385280449583919, "grad_norm": 0.4516046345233917, "learning_rate": 8e-05, "loss": 1.8799, "step": 4983 }, { "epoch": 0.5386361180157787, "grad_norm": 0.39778828620910645, "learning_rate": 8e-05, "loss": 1.7203, "step": 4984 }, { "epoch": 0.5387441910731655, "grad_norm": 0.41983702778816223, "learning_rate": 8e-05, "loss": 1.6829, "step": 4985 }, { "epoch": 0.5388522641305522, "grad_norm": 0.39821866154670715, "learning_rate": 8e-05, "loss": 1.7242, "step": 4986 }, { "epoch": 0.538960337187939, "grad_norm": 0.3996838629245758, "learning_rate": 8e-05, "loss": 1.816, "step": 4987 }, { "epoch": 0.5390684102453258, "grad_norm": 0.41400259733200073, "learning_rate": 8e-05, "loss": 1.6538, "step": 4988 }, { "epoch": 0.5391764833027126, "grad_norm": 0.4749830961227417, "learning_rate": 8e-05, "loss": 1.9272, "step": 4989 }, { "epoch": 0.5392845563600994, "grad_norm": 0.40028098225593567, "learning_rate": 8e-05, "loss": 1.5645, "step": 4990 }, { "epoch": 0.5393926294174862, "grad_norm": 0.4243418872356415, "learning_rate": 8e-05, "loss": 1.7819, "step": 4991 }, { "epoch": 0.539500702474873, "grad_norm": 0.42156633734703064, "learning_rate": 8e-05, "loss": 1.7393, "step": 4992 }, { "epoch": 0.5396087755322598, "grad_norm": 0.4170660078525543, "learning_rate": 8e-05, "loss": 1.8224, "step": 4993 }, { "epoch": 0.5397168485896466, "grad_norm": 0.4410928189754486, "learning_rate": 8e-05, "loss": 1.8482, "step": 4994 }, { "epoch": 0.5398249216470334, "grad_norm": 0.45637479424476624, "learning_rate": 8e-05, "loss": 1.9589, "step": 4995 }, { "epoch": 0.5399329947044201, "grad_norm": 0.3825331926345825, "learning_rate": 8e-05, "loss": 1.5749, "step": 4996 }, { "epoch": 0.5400410677618069, "grad_norm": 0.4709583818912506, "learning_rate": 8e-05, "loss": 1.8035, "step": 4997 }, { "epoch": 0.5401491408191937, "grad_norm": 0.44993558526039124, "learning_rate": 8e-05, "loss": 1.8077, "step": 4998 }, { "epoch": 0.5402572138765805, "grad_norm": 0.44802701473236084, "learning_rate": 8e-05, "loss": 1.6869, "step": 4999 }, { "epoch": 0.5403652869339673, "grad_norm": 0.4343043267726898, "learning_rate": 8e-05, "loss": 1.739, "step": 5000 }, { "epoch": 0.5404733599913542, "grad_norm": 0.4244413673877716, "learning_rate": 8e-05, "loss": 1.922, "step": 5001 }, { "epoch": 0.540581433048741, "grad_norm": 0.41614779829978943, "learning_rate": 8e-05, "loss": 1.6699, "step": 5002 }, { "epoch": 0.5406895061061278, "grad_norm": 0.37839698791503906, "learning_rate": 8e-05, "loss": 1.6446, "step": 5003 }, { "epoch": 0.5407975791635146, "grad_norm": 0.3808358311653137, "learning_rate": 8e-05, "loss": 1.6094, "step": 5004 }, { "epoch": 0.5409056522209014, "grad_norm": 0.4267764687538147, "learning_rate": 8e-05, "loss": 1.6253, "step": 5005 }, { "epoch": 0.5410137252782882, "grad_norm": 0.4169236421585083, "learning_rate": 8e-05, "loss": 1.7009, "step": 5006 }, { "epoch": 0.5411217983356749, "grad_norm": 0.40359020233154297, "learning_rate": 8e-05, "loss": 1.6213, "step": 5007 }, { "epoch": 0.5412298713930617, "grad_norm": 0.42857298254966736, "learning_rate": 8e-05, "loss": 1.6256, "step": 5008 }, { "epoch": 0.5413379444504485, "grad_norm": 0.42052945494651794, "learning_rate": 8e-05, "loss": 1.7457, "step": 5009 }, { "epoch": 0.5414460175078353, "grad_norm": 0.3793718218803406, "learning_rate": 8e-05, "loss": 1.7383, "step": 5010 }, { "epoch": 0.5415540905652221, "grad_norm": 0.41419556736946106, "learning_rate": 8e-05, "loss": 1.7065, "step": 5011 }, { "epoch": 0.5416621636226089, "grad_norm": 0.3876740336418152, "learning_rate": 8e-05, "loss": 1.615, "step": 5012 }, { "epoch": 0.5417702366799957, "grad_norm": 0.40463772416114807, "learning_rate": 8e-05, "loss": 1.6124, "step": 5013 }, { "epoch": 0.5418783097373825, "grad_norm": 0.4117019474506378, "learning_rate": 8e-05, "loss": 1.5995, "step": 5014 }, { "epoch": 0.5419863827947693, "grad_norm": 0.4699041545391083, "learning_rate": 8e-05, "loss": 1.7574, "step": 5015 }, { "epoch": 0.5420944558521561, "grad_norm": 0.4097781777381897, "learning_rate": 8e-05, "loss": 1.6871, "step": 5016 }, { "epoch": 0.5422025289095429, "grad_norm": 0.4295152425765991, "learning_rate": 8e-05, "loss": 1.8496, "step": 5017 }, { "epoch": 0.5423106019669296, "grad_norm": 0.410518079996109, "learning_rate": 8e-05, "loss": 1.695, "step": 5018 }, { "epoch": 0.5424186750243164, "grad_norm": 0.4101126492023468, "learning_rate": 8e-05, "loss": 1.9252, "step": 5019 }, { "epoch": 0.5425267480817032, "grad_norm": 0.4521569013595581, "learning_rate": 8e-05, "loss": 1.9035, "step": 5020 }, { "epoch": 0.54263482113909, "grad_norm": 0.4086431562900543, "learning_rate": 8e-05, "loss": 1.6917, "step": 5021 }, { "epoch": 0.5427428941964768, "grad_norm": 0.4508405327796936, "learning_rate": 8e-05, "loss": 1.965, "step": 5022 }, { "epoch": 0.5428509672538636, "grad_norm": 0.40451905131340027, "learning_rate": 8e-05, "loss": 1.6986, "step": 5023 }, { "epoch": 0.5429590403112504, "grad_norm": 0.4337192177772522, "learning_rate": 8e-05, "loss": 1.808, "step": 5024 }, { "epoch": 0.5430671133686372, "grad_norm": 0.38895702362060547, "learning_rate": 8e-05, "loss": 1.6736, "step": 5025 }, { "epoch": 0.543175186426024, "grad_norm": 0.41385072469711304, "learning_rate": 8e-05, "loss": 1.7581, "step": 5026 }, { "epoch": 0.5432832594834108, "grad_norm": 0.3876224458217621, "learning_rate": 8e-05, "loss": 1.6112, "step": 5027 }, { "epoch": 0.5433913325407976, "grad_norm": 0.4359486401081085, "learning_rate": 8e-05, "loss": 1.7855, "step": 5028 }, { "epoch": 0.5434994055981843, "grad_norm": 0.37843725085258484, "learning_rate": 8e-05, "loss": 1.6537, "step": 5029 }, { "epoch": 0.5436074786555711, "grad_norm": 0.41212618350982666, "learning_rate": 8e-05, "loss": 1.5656, "step": 5030 }, { "epoch": 0.5437155517129579, "grad_norm": 0.4118042588233948, "learning_rate": 8e-05, "loss": 1.7524, "step": 5031 }, { "epoch": 0.5438236247703447, "grad_norm": 0.42473241686820984, "learning_rate": 8e-05, "loss": 1.8095, "step": 5032 }, { "epoch": 0.5439316978277315, "grad_norm": 0.39658576250076294, "learning_rate": 8e-05, "loss": 1.7776, "step": 5033 }, { "epoch": 0.5440397708851183, "grad_norm": 0.4162738025188446, "learning_rate": 8e-05, "loss": 1.7594, "step": 5034 }, { "epoch": 0.5441478439425051, "grad_norm": 0.3970663845539093, "learning_rate": 8e-05, "loss": 1.7439, "step": 5035 }, { "epoch": 0.544255916999892, "grad_norm": 0.43537357449531555, "learning_rate": 8e-05, "loss": 1.8108, "step": 5036 }, { "epoch": 0.5443639900572788, "grad_norm": 0.46973299980163574, "learning_rate": 8e-05, "loss": 1.7683, "step": 5037 }, { "epoch": 0.5444720631146656, "grad_norm": 0.40190377831459045, "learning_rate": 8e-05, "loss": 1.6934, "step": 5038 }, { "epoch": 0.5445801361720523, "grad_norm": 0.3977755606174469, "learning_rate": 8e-05, "loss": 1.6014, "step": 5039 }, { "epoch": 0.5446882092294391, "grad_norm": 0.45302242040634155, "learning_rate": 8e-05, "loss": 1.8084, "step": 5040 }, { "epoch": 0.5447962822868259, "grad_norm": 0.4094148278236389, "learning_rate": 8e-05, "loss": 1.7896, "step": 5041 }, { "epoch": 0.5449043553442127, "grad_norm": 0.37804582715034485, "learning_rate": 8e-05, "loss": 1.5816, "step": 5042 }, { "epoch": 0.5450124284015995, "grad_norm": 0.47089770436286926, "learning_rate": 8e-05, "loss": 1.7745, "step": 5043 }, { "epoch": 0.5451205014589863, "grad_norm": 0.43440115451812744, "learning_rate": 8e-05, "loss": 1.731, "step": 5044 }, { "epoch": 0.5452285745163731, "grad_norm": 0.3920232057571411, "learning_rate": 8e-05, "loss": 1.5698, "step": 5045 }, { "epoch": 0.5453366475737599, "grad_norm": 0.3765890598297119, "learning_rate": 8e-05, "loss": 1.6178, "step": 5046 }, { "epoch": 0.5454447206311467, "grad_norm": 0.4566970467567444, "learning_rate": 8e-05, "loss": 1.8671, "step": 5047 }, { "epoch": 0.5455527936885335, "grad_norm": 0.4800170063972473, "learning_rate": 8e-05, "loss": 1.7469, "step": 5048 }, { "epoch": 0.5456608667459203, "grad_norm": 0.44118955731391907, "learning_rate": 8e-05, "loss": 1.6051, "step": 5049 }, { "epoch": 0.545768939803307, "grad_norm": 0.4074260890483856, "learning_rate": 8e-05, "loss": 1.6206, "step": 5050 }, { "epoch": 0.5458770128606938, "grad_norm": 0.3939186930656433, "learning_rate": 8e-05, "loss": 1.6831, "step": 5051 }, { "epoch": 0.5459850859180806, "grad_norm": 0.40486904978752136, "learning_rate": 8e-05, "loss": 1.6358, "step": 5052 }, { "epoch": 0.5460931589754674, "grad_norm": 0.443316787481308, "learning_rate": 8e-05, "loss": 1.6273, "step": 5053 }, { "epoch": 0.5462012320328542, "grad_norm": 0.39102137088775635, "learning_rate": 8e-05, "loss": 1.5602, "step": 5054 }, { "epoch": 0.546309305090241, "grad_norm": 0.39311760663986206, "learning_rate": 8e-05, "loss": 1.5818, "step": 5055 }, { "epoch": 0.5464173781476278, "grad_norm": 0.3975493311882019, "learning_rate": 8e-05, "loss": 1.6726, "step": 5056 }, { "epoch": 0.5465254512050146, "grad_norm": 0.448491632938385, "learning_rate": 8e-05, "loss": 1.9616, "step": 5057 }, { "epoch": 0.5466335242624014, "grad_norm": 0.39220985770225525, "learning_rate": 8e-05, "loss": 1.6561, "step": 5058 }, { "epoch": 0.5467415973197882, "grad_norm": 0.4087098240852356, "learning_rate": 8e-05, "loss": 1.5872, "step": 5059 }, { "epoch": 0.546849670377175, "grad_norm": 0.4325299561023712, "learning_rate": 8e-05, "loss": 1.7826, "step": 5060 }, { "epoch": 0.5469577434345617, "grad_norm": 0.4423466622829437, "learning_rate": 8e-05, "loss": 1.7491, "step": 5061 }, { "epoch": 0.5470658164919485, "grad_norm": 0.42137977480888367, "learning_rate": 8e-05, "loss": 1.5364, "step": 5062 }, { "epoch": 0.5471738895493353, "grad_norm": 0.40570077300071716, "learning_rate": 8e-05, "loss": 1.633, "step": 5063 }, { "epoch": 0.5472819626067221, "grad_norm": 0.39775902032852173, "learning_rate": 8e-05, "loss": 1.5698, "step": 5064 }, { "epoch": 0.5473900356641089, "grad_norm": 0.4204985201358795, "learning_rate": 8e-05, "loss": 1.7193, "step": 5065 }, { "epoch": 0.5474981087214957, "grad_norm": 0.39921531081199646, "learning_rate": 8e-05, "loss": 1.5966, "step": 5066 }, { "epoch": 0.5476061817788825, "grad_norm": 0.3889120817184448, "learning_rate": 8e-05, "loss": 1.4955, "step": 5067 }, { "epoch": 0.5477142548362693, "grad_norm": 0.40015801787376404, "learning_rate": 8e-05, "loss": 1.7547, "step": 5068 }, { "epoch": 0.5478223278936561, "grad_norm": 0.41209420561790466, "learning_rate": 8e-05, "loss": 1.7058, "step": 5069 }, { "epoch": 0.547930400951043, "grad_norm": 0.4043588936328888, "learning_rate": 8e-05, "loss": 1.5726, "step": 5070 }, { "epoch": 0.5480384740084296, "grad_norm": 0.3962462842464447, "learning_rate": 8e-05, "loss": 1.6975, "step": 5071 }, { "epoch": 0.5481465470658164, "grad_norm": 0.42182236909866333, "learning_rate": 8e-05, "loss": 1.7734, "step": 5072 }, { "epoch": 0.5482546201232033, "grad_norm": 0.4000571370124817, "learning_rate": 8e-05, "loss": 1.7688, "step": 5073 }, { "epoch": 0.54836269318059, "grad_norm": 0.38760972023010254, "learning_rate": 8e-05, "loss": 1.4282, "step": 5074 }, { "epoch": 0.5484707662379769, "grad_norm": 0.46368008852005005, "learning_rate": 8e-05, "loss": 1.8831, "step": 5075 }, { "epoch": 0.5485788392953637, "grad_norm": 0.40976059436798096, "learning_rate": 8e-05, "loss": 1.6657, "step": 5076 }, { "epoch": 0.5486869123527505, "grad_norm": 0.3994709253311157, "learning_rate": 8e-05, "loss": 1.5157, "step": 5077 }, { "epoch": 0.5487949854101373, "grad_norm": 0.4202108383178711, "learning_rate": 8e-05, "loss": 1.5151, "step": 5078 }, { "epoch": 0.5489030584675241, "grad_norm": 0.40741753578186035, "learning_rate": 8e-05, "loss": 1.6449, "step": 5079 }, { "epoch": 0.5490111315249109, "grad_norm": 0.389801949262619, "learning_rate": 8e-05, "loss": 1.6378, "step": 5080 }, { "epoch": 0.5491192045822977, "grad_norm": 0.4083135724067688, "learning_rate": 8e-05, "loss": 1.572, "step": 5081 }, { "epoch": 0.5492272776396844, "grad_norm": 0.43986016511917114, "learning_rate": 8e-05, "loss": 1.688, "step": 5082 }, { "epoch": 0.5493353506970712, "grad_norm": 0.3950330317020416, "learning_rate": 8e-05, "loss": 1.576, "step": 5083 }, { "epoch": 0.549443423754458, "grad_norm": 0.41498368978500366, "learning_rate": 8e-05, "loss": 1.6997, "step": 5084 }, { "epoch": 0.5495514968118448, "grad_norm": 0.44374847412109375, "learning_rate": 8e-05, "loss": 1.6771, "step": 5085 }, { "epoch": 0.5496595698692316, "grad_norm": 0.4404027462005615, "learning_rate": 8e-05, "loss": 1.8412, "step": 5086 }, { "epoch": 0.5497676429266184, "grad_norm": 0.42734891176223755, "learning_rate": 8e-05, "loss": 1.7078, "step": 5087 }, { "epoch": 0.5498757159840052, "grad_norm": 0.3955828845500946, "learning_rate": 8e-05, "loss": 1.6094, "step": 5088 }, { "epoch": 0.549983789041392, "grad_norm": 0.4362671673297882, "learning_rate": 8e-05, "loss": 1.7875, "step": 5089 }, { "epoch": 0.5500918620987788, "grad_norm": 0.3982241451740265, "learning_rate": 8e-05, "loss": 1.7192, "step": 5090 }, { "epoch": 0.5501999351561656, "grad_norm": 0.3938392102718353, "learning_rate": 8e-05, "loss": 1.5863, "step": 5091 }, { "epoch": 0.5503080082135524, "grad_norm": 0.4448455274105072, "learning_rate": 8e-05, "loss": 1.8592, "step": 5092 }, { "epoch": 0.5504160812709391, "grad_norm": 0.4748949408531189, "learning_rate": 8e-05, "loss": 1.6739, "step": 5093 }, { "epoch": 0.5505241543283259, "grad_norm": 0.4705922305583954, "learning_rate": 8e-05, "loss": 1.9184, "step": 5094 }, { "epoch": 0.5506322273857127, "grad_norm": 0.44562968611717224, "learning_rate": 8e-05, "loss": 1.7197, "step": 5095 }, { "epoch": 0.5507403004430995, "grad_norm": 0.4073898494243622, "learning_rate": 8e-05, "loss": 1.7031, "step": 5096 }, { "epoch": 0.5508483735004863, "grad_norm": 0.4443080723285675, "learning_rate": 8e-05, "loss": 1.8101, "step": 5097 }, { "epoch": 0.5509564465578731, "grad_norm": 0.402456670999527, "learning_rate": 8e-05, "loss": 1.657, "step": 5098 }, { "epoch": 0.5510645196152599, "grad_norm": 0.4095904529094696, "learning_rate": 8e-05, "loss": 1.7167, "step": 5099 }, { "epoch": 0.5511725926726467, "grad_norm": 0.4191281795501709, "learning_rate": 8e-05, "loss": 1.632, "step": 5100 }, { "epoch": 0.5512806657300335, "grad_norm": 0.4488406181335449, "learning_rate": 8e-05, "loss": 1.6579, "step": 5101 }, { "epoch": 0.5513887387874203, "grad_norm": 0.4296759068965912, "learning_rate": 8e-05, "loss": 1.6722, "step": 5102 }, { "epoch": 0.5514968118448071, "grad_norm": 0.3852338194847107, "learning_rate": 8e-05, "loss": 1.611, "step": 5103 }, { "epoch": 0.5516048849021938, "grad_norm": 0.3966812193393707, "learning_rate": 8e-05, "loss": 1.5931, "step": 5104 }, { "epoch": 0.5517129579595806, "grad_norm": 0.41980767250061035, "learning_rate": 8e-05, "loss": 1.7068, "step": 5105 }, { "epoch": 0.5518210310169674, "grad_norm": 0.4188380539417267, "learning_rate": 8e-05, "loss": 1.7147, "step": 5106 }, { "epoch": 0.5519291040743542, "grad_norm": 0.43606263399124146, "learning_rate": 8e-05, "loss": 1.8017, "step": 5107 }, { "epoch": 0.552037177131741, "grad_norm": 0.4357244372367859, "learning_rate": 8e-05, "loss": 1.7995, "step": 5108 }, { "epoch": 0.5521452501891279, "grad_norm": 0.40860825777053833, "learning_rate": 8e-05, "loss": 1.664, "step": 5109 }, { "epoch": 0.5522533232465147, "grad_norm": 0.3938402533531189, "learning_rate": 8e-05, "loss": 1.7545, "step": 5110 }, { "epoch": 0.5523613963039015, "grad_norm": 0.37914443016052246, "learning_rate": 8e-05, "loss": 1.6283, "step": 5111 }, { "epoch": 0.5524694693612883, "grad_norm": 0.4374390244483948, "learning_rate": 8e-05, "loss": 1.7658, "step": 5112 }, { "epoch": 0.5525775424186751, "grad_norm": 0.4690524935722351, "learning_rate": 8e-05, "loss": 1.8776, "step": 5113 }, { "epoch": 0.5526856154760618, "grad_norm": 0.42699718475341797, "learning_rate": 8e-05, "loss": 1.8182, "step": 5114 }, { "epoch": 0.5527936885334486, "grad_norm": 0.41284996271133423, "learning_rate": 8e-05, "loss": 1.7308, "step": 5115 }, { "epoch": 0.5529017615908354, "grad_norm": 0.382125586271286, "learning_rate": 8e-05, "loss": 1.6765, "step": 5116 }, { "epoch": 0.5530098346482222, "grad_norm": 0.39979568123817444, "learning_rate": 8e-05, "loss": 1.3289, "step": 5117 }, { "epoch": 0.553117907705609, "grad_norm": 0.4564915895462036, "learning_rate": 8e-05, "loss": 1.8273, "step": 5118 }, { "epoch": 0.5532259807629958, "grad_norm": 0.44492167234420776, "learning_rate": 8e-05, "loss": 1.7912, "step": 5119 }, { "epoch": 0.5533340538203826, "grad_norm": 0.41281694173812866, "learning_rate": 8e-05, "loss": 1.6697, "step": 5120 }, { "epoch": 0.5534421268777694, "grad_norm": 0.40094250440597534, "learning_rate": 8e-05, "loss": 1.6774, "step": 5121 }, { "epoch": 0.5535501999351562, "grad_norm": 0.47668859362602234, "learning_rate": 8e-05, "loss": 1.9167, "step": 5122 }, { "epoch": 0.553658272992543, "grad_norm": 0.5074390172958374, "learning_rate": 8e-05, "loss": 1.7928, "step": 5123 }, { "epoch": 0.5537663460499298, "grad_norm": 0.4405800998210907, "learning_rate": 8e-05, "loss": 1.6557, "step": 5124 }, { "epoch": 0.5538744191073165, "grad_norm": 0.4306829273700714, "learning_rate": 8e-05, "loss": 1.8884, "step": 5125 }, { "epoch": 0.5539824921647033, "grad_norm": 0.40004733204841614, "learning_rate": 8e-05, "loss": 1.5353, "step": 5126 }, { "epoch": 0.5540905652220901, "grad_norm": 0.42598506808280945, "learning_rate": 8e-05, "loss": 1.4422, "step": 5127 }, { "epoch": 0.5541986382794769, "grad_norm": 0.4495410919189453, "learning_rate": 8e-05, "loss": 1.937, "step": 5128 }, { "epoch": 0.5543067113368637, "grad_norm": 0.4180264174938202, "learning_rate": 8e-05, "loss": 1.7936, "step": 5129 }, { "epoch": 0.5544147843942505, "grad_norm": 0.41504359245300293, "learning_rate": 8e-05, "loss": 1.7061, "step": 5130 }, { "epoch": 0.5545228574516373, "grad_norm": 0.40480008721351624, "learning_rate": 8e-05, "loss": 1.849, "step": 5131 }, { "epoch": 0.5546309305090241, "grad_norm": 0.41416826844215393, "learning_rate": 8e-05, "loss": 1.7693, "step": 5132 }, { "epoch": 0.5547390035664109, "grad_norm": 0.4087192714214325, "learning_rate": 8e-05, "loss": 1.6934, "step": 5133 }, { "epoch": 0.5548470766237977, "grad_norm": 0.40085700154304504, "learning_rate": 8e-05, "loss": 1.6625, "step": 5134 }, { "epoch": 0.5549551496811845, "grad_norm": 0.40817028284072876, "learning_rate": 8e-05, "loss": 1.7282, "step": 5135 }, { "epoch": 0.5550632227385712, "grad_norm": 0.44116246700286865, "learning_rate": 8e-05, "loss": 1.7626, "step": 5136 }, { "epoch": 0.555171295795958, "grad_norm": 0.4251325726509094, "learning_rate": 8e-05, "loss": 1.7908, "step": 5137 }, { "epoch": 0.5552793688533448, "grad_norm": 0.3858116567134857, "learning_rate": 8e-05, "loss": 1.5886, "step": 5138 }, { "epoch": 0.5553874419107316, "grad_norm": 0.48880621790885925, "learning_rate": 8e-05, "loss": 1.8382, "step": 5139 }, { "epoch": 0.5554955149681184, "grad_norm": 0.40059101581573486, "learning_rate": 8e-05, "loss": 1.8208, "step": 5140 }, { "epoch": 0.5556035880255052, "grad_norm": 0.4540790617465973, "learning_rate": 8e-05, "loss": 1.8741, "step": 5141 }, { "epoch": 0.555711661082892, "grad_norm": 0.3981870412826538, "learning_rate": 8e-05, "loss": 1.728, "step": 5142 }, { "epoch": 0.5558197341402789, "grad_norm": 0.434050977230072, "learning_rate": 8e-05, "loss": 1.6438, "step": 5143 }, { "epoch": 0.5559278071976657, "grad_norm": 0.4042028784751892, "learning_rate": 8e-05, "loss": 1.5687, "step": 5144 }, { "epoch": 0.5560358802550525, "grad_norm": 0.4119265079498291, "learning_rate": 8e-05, "loss": 1.7555, "step": 5145 }, { "epoch": 0.5561439533124392, "grad_norm": 0.3891986608505249, "learning_rate": 8e-05, "loss": 1.4762, "step": 5146 }, { "epoch": 0.556252026369826, "grad_norm": 0.454012006521225, "learning_rate": 8e-05, "loss": 1.5842, "step": 5147 }, { "epoch": 0.5563600994272128, "grad_norm": 0.4255113899707794, "learning_rate": 8e-05, "loss": 1.8156, "step": 5148 }, { "epoch": 0.5564681724845996, "grad_norm": 0.4092005789279938, "learning_rate": 8e-05, "loss": 1.6759, "step": 5149 }, { "epoch": 0.5565762455419864, "grad_norm": 0.4649369716644287, "learning_rate": 8e-05, "loss": 1.8533, "step": 5150 }, { "epoch": 0.5566843185993732, "grad_norm": 0.4249788522720337, "learning_rate": 8e-05, "loss": 1.7825, "step": 5151 }, { "epoch": 0.55679239165676, "grad_norm": 0.4163335859775543, "learning_rate": 8e-05, "loss": 1.8636, "step": 5152 }, { "epoch": 0.5569004647141468, "grad_norm": 0.4123961329460144, "learning_rate": 8e-05, "loss": 1.6593, "step": 5153 }, { "epoch": 0.5570085377715336, "grad_norm": 0.426469624042511, "learning_rate": 8e-05, "loss": 1.6317, "step": 5154 }, { "epoch": 0.5571166108289204, "grad_norm": 0.46177300810813904, "learning_rate": 8e-05, "loss": 1.7549, "step": 5155 }, { "epoch": 0.5572246838863072, "grad_norm": 0.4006745517253876, "learning_rate": 8e-05, "loss": 1.6204, "step": 5156 }, { "epoch": 0.5573327569436939, "grad_norm": 0.40350422263145447, "learning_rate": 8e-05, "loss": 1.6183, "step": 5157 }, { "epoch": 0.5574408300010807, "grad_norm": 0.3892279863357544, "learning_rate": 8e-05, "loss": 1.7373, "step": 5158 }, { "epoch": 0.5575489030584675, "grad_norm": 0.38431432843208313, "learning_rate": 8e-05, "loss": 1.6676, "step": 5159 }, { "epoch": 0.5576569761158543, "grad_norm": 0.4244755804538727, "learning_rate": 8e-05, "loss": 1.718, "step": 5160 }, { "epoch": 0.5577650491732411, "grad_norm": 0.4135211408138275, "learning_rate": 8e-05, "loss": 1.7156, "step": 5161 }, { "epoch": 0.5578731222306279, "grad_norm": 0.39573052525520325, "learning_rate": 8e-05, "loss": 1.609, "step": 5162 }, { "epoch": 0.5579811952880147, "grad_norm": 0.3862548768520355, "learning_rate": 8e-05, "loss": 1.5558, "step": 5163 }, { "epoch": 0.5580892683454015, "grad_norm": 0.39231184124946594, "learning_rate": 8e-05, "loss": 1.6416, "step": 5164 }, { "epoch": 0.5581973414027883, "grad_norm": 0.394728422164917, "learning_rate": 8e-05, "loss": 1.6331, "step": 5165 }, { "epoch": 0.5583054144601751, "grad_norm": 0.41758570075035095, "learning_rate": 8e-05, "loss": 1.6189, "step": 5166 }, { "epoch": 0.5584134875175619, "grad_norm": 0.435490220785141, "learning_rate": 8e-05, "loss": 1.8504, "step": 5167 }, { "epoch": 0.5585215605749486, "grad_norm": 0.4611351788043976, "learning_rate": 8e-05, "loss": 1.9141, "step": 5168 }, { "epoch": 0.5586296336323354, "grad_norm": 0.4218725264072418, "learning_rate": 8e-05, "loss": 1.7411, "step": 5169 }, { "epoch": 0.5587377066897222, "grad_norm": 0.42524969577789307, "learning_rate": 8e-05, "loss": 1.7247, "step": 5170 }, { "epoch": 0.558845779747109, "grad_norm": 0.3805936276912689, "learning_rate": 8e-05, "loss": 1.7459, "step": 5171 }, { "epoch": 0.5589538528044958, "grad_norm": 0.3879164755344391, "learning_rate": 8e-05, "loss": 1.6541, "step": 5172 }, { "epoch": 0.5590619258618826, "grad_norm": 0.42860686779022217, "learning_rate": 8e-05, "loss": 1.5732, "step": 5173 }, { "epoch": 0.5591699989192694, "grad_norm": 0.4078417122364044, "learning_rate": 8e-05, "loss": 1.7587, "step": 5174 }, { "epoch": 0.5592780719766562, "grad_norm": 0.388156920671463, "learning_rate": 8e-05, "loss": 1.6631, "step": 5175 }, { "epoch": 0.559386145034043, "grad_norm": 0.41122227907180786, "learning_rate": 8e-05, "loss": 1.7642, "step": 5176 }, { "epoch": 0.5594942180914299, "grad_norm": 0.39018478989601135, "learning_rate": 8e-05, "loss": 1.6048, "step": 5177 }, { "epoch": 0.5596022911488167, "grad_norm": 0.4558615982532501, "learning_rate": 8e-05, "loss": 1.852, "step": 5178 }, { "epoch": 0.5597103642062033, "grad_norm": 0.40256115794181824, "learning_rate": 8e-05, "loss": 1.6221, "step": 5179 }, { "epoch": 0.5598184372635902, "grad_norm": 0.4127877950668335, "learning_rate": 8e-05, "loss": 1.6127, "step": 5180 }, { "epoch": 0.559926510320977, "grad_norm": 0.4073278307914734, "learning_rate": 8e-05, "loss": 1.7627, "step": 5181 }, { "epoch": 0.5600345833783638, "grad_norm": 0.456869900226593, "learning_rate": 8e-05, "loss": 1.6409, "step": 5182 }, { "epoch": 0.5601426564357506, "grad_norm": 0.4028075635433197, "learning_rate": 8e-05, "loss": 1.6952, "step": 5183 }, { "epoch": 0.5602507294931374, "grad_norm": 0.46084922552108765, "learning_rate": 8e-05, "loss": 1.9426, "step": 5184 }, { "epoch": 0.5603588025505242, "grad_norm": 0.42392438650131226, "learning_rate": 8e-05, "loss": 1.8531, "step": 5185 }, { "epoch": 0.560466875607911, "grad_norm": 0.3893096446990967, "learning_rate": 8e-05, "loss": 1.7308, "step": 5186 }, { "epoch": 0.5605749486652978, "grad_norm": 0.40702494978904724, "learning_rate": 8e-05, "loss": 1.7567, "step": 5187 }, { "epoch": 0.5606830217226846, "grad_norm": 0.4268399178981781, "learning_rate": 8e-05, "loss": 1.8795, "step": 5188 }, { "epoch": 0.5607910947800713, "grad_norm": 0.4087476134300232, "learning_rate": 8e-05, "loss": 1.6466, "step": 5189 }, { "epoch": 0.5608991678374581, "grad_norm": 0.4119463264942169, "learning_rate": 8e-05, "loss": 1.7566, "step": 5190 }, { "epoch": 0.5610072408948449, "grad_norm": 0.4584994912147522, "learning_rate": 8e-05, "loss": 1.8651, "step": 5191 }, { "epoch": 0.5611153139522317, "grad_norm": 0.48046889901161194, "learning_rate": 8e-05, "loss": 1.6654, "step": 5192 }, { "epoch": 0.5612233870096185, "grad_norm": 0.4189452528953552, "learning_rate": 8e-05, "loss": 1.7278, "step": 5193 }, { "epoch": 0.5613314600670053, "grad_norm": 0.463471382856369, "learning_rate": 8e-05, "loss": 1.8386, "step": 5194 }, { "epoch": 0.5614395331243921, "grad_norm": 0.3817075490951538, "learning_rate": 8e-05, "loss": 1.537, "step": 5195 }, { "epoch": 0.5615476061817789, "grad_norm": 0.3908224403858185, "learning_rate": 8e-05, "loss": 1.5096, "step": 5196 }, { "epoch": 0.5616556792391657, "grad_norm": 0.41907885670661926, "learning_rate": 8e-05, "loss": 1.7639, "step": 5197 }, { "epoch": 0.5617637522965525, "grad_norm": 0.403381884098053, "learning_rate": 8e-05, "loss": 1.6568, "step": 5198 }, { "epoch": 0.5618718253539393, "grad_norm": 0.4285737872123718, "learning_rate": 8e-05, "loss": 1.7847, "step": 5199 }, { "epoch": 0.561979898411326, "grad_norm": 0.4595038592815399, "learning_rate": 8e-05, "loss": 1.695, "step": 5200 }, { "epoch": 0.5620879714687128, "grad_norm": 0.4161752164363861, "learning_rate": 8e-05, "loss": 1.8569, "step": 5201 }, { "epoch": 0.5621960445260996, "grad_norm": 0.3962073028087616, "learning_rate": 8e-05, "loss": 1.606, "step": 5202 }, { "epoch": 0.5623041175834864, "grad_norm": 0.3813258409500122, "learning_rate": 8e-05, "loss": 1.5785, "step": 5203 }, { "epoch": 0.5624121906408732, "grad_norm": 0.41440603137016296, "learning_rate": 8e-05, "loss": 1.6676, "step": 5204 }, { "epoch": 0.56252026369826, "grad_norm": 0.41403743624687195, "learning_rate": 8e-05, "loss": 1.9024, "step": 5205 }, { "epoch": 0.5626283367556468, "grad_norm": 0.41665706038475037, "learning_rate": 8e-05, "loss": 1.7847, "step": 5206 }, { "epoch": 0.5627364098130336, "grad_norm": 0.4062924087047577, "learning_rate": 8e-05, "loss": 1.7042, "step": 5207 }, { "epoch": 0.5628444828704204, "grad_norm": 0.38252437114715576, "learning_rate": 8e-05, "loss": 1.5884, "step": 5208 }, { "epoch": 0.5629525559278072, "grad_norm": 0.4042649567127228, "learning_rate": 8e-05, "loss": 1.7366, "step": 5209 }, { "epoch": 0.563060628985194, "grad_norm": 0.43591606616973877, "learning_rate": 8e-05, "loss": 1.6739, "step": 5210 }, { "epoch": 0.5631687020425807, "grad_norm": 0.45777586102485657, "learning_rate": 8e-05, "loss": 1.8482, "step": 5211 }, { "epoch": 0.5632767750999675, "grad_norm": 0.423603892326355, "learning_rate": 8e-05, "loss": 1.9179, "step": 5212 }, { "epoch": 0.5633848481573543, "grad_norm": 0.39726337790489197, "learning_rate": 8e-05, "loss": 1.7153, "step": 5213 }, { "epoch": 0.5634929212147411, "grad_norm": 0.4041389524936676, "learning_rate": 8e-05, "loss": 1.7127, "step": 5214 }, { "epoch": 0.563600994272128, "grad_norm": 0.44530096650123596, "learning_rate": 8e-05, "loss": 1.8482, "step": 5215 }, { "epoch": 0.5637090673295148, "grad_norm": 0.45409882068634033, "learning_rate": 8e-05, "loss": 1.6581, "step": 5216 }, { "epoch": 0.5638171403869016, "grad_norm": 0.4372636675834656, "learning_rate": 8e-05, "loss": 1.6872, "step": 5217 }, { "epoch": 0.5639252134442884, "grad_norm": 0.3993006646633148, "learning_rate": 8e-05, "loss": 1.6789, "step": 5218 }, { "epoch": 0.5640332865016752, "grad_norm": 0.4224236309528351, "learning_rate": 8e-05, "loss": 1.8013, "step": 5219 }, { "epoch": 0.564141359559062, "grad_norm": 0.41767793893814087, "learning_rate": 8e-05, "loss": 1.7169, "step": 5220 }, { "epoch": 0.5642494326164487, "grad_norm": 0.4484507143497467, "learning_rate": 8e-05, "loss": 1.9581, "step": 5221 }, { "epoch": 0.5643575056738355, "grad_norm": 0.4190903902053833, "learning_rate": 8e-05, "loss": 1.8223, "step": 5222 }, { "epoch": 0.5644655787312223, "grad_norm": 0.41674375534057617, "learning_rate": 8e-05, "loss": 1.7078, "step": 5223 }, { "epoch": 0.5645736517886091, "grad_norm": 0.41433200240135193, "learning_rate": 8e-05, "loss": 1.7301, "step": 5224 }, { "epoch": 0.5646817248459959, "grad_norm": 0.42008456587791443, "learning_rate": 8e-05, "loss": 1.7863, "step": 5225 }, { "epoch": 0.5647897979033827, "grad_norm": 0.4315822422504425, "learning_rate": 8e-05, "loss": 1.8509, "step": 5226 }, { "epoch": 0.5648978709607695, "grad_norm": 0.371372252702713, "learning_rate": 8e-05, "loss": 1.6121, "step": 5227 }, { "epoch": 0.5650059440181563, "grad_norm": 0.40731194615364075, "learning_rate": 8e-05, "loss": 1.672, "step": 5228 }, { "epoch": 0.5651140170755431, "grad_norm": 0.4152587056159973, "learning_rate": 8e-05, "loss": 1.807, "step": 5229 }, { "epoch": 0.5652220901329299, "grad_norm": 0.39882731437683105, "learning_rate": 8e-05, "loss": 1.7742, "step": 5230 }, { "epoch": 0.5653301631903167, "grad_norm": 0.4301312267780304, "learning_rate": 8e-05, "loss": 1.7382, "step": 5231 }, { "epoch": 0.5654382362477034, "grad_norm": 0.39597368240356445, "learning_rate": 8e-05, "loss": 1.5683, "step": 5232 }, { "epoch": 0.5655463093050902, "grad_norm": 0.3978910744190216, "learning_rate": 8e-05, "loss": 1.602, "step": 5233 }, { "epoch": 0.565654382362477, "grad_norm": 0.4744794964790344, "learning_rate": 8e-05, "loss": 1.8997, "step": 5234 }, { "epoch": 0.5657624554198638, "grad_norm": 0.44149982929229736, "learning_rate": 8e-05, "loss": 1.787, "step": 5235 }, { "epoch": 0.5658705284772506, "grad_norm": 0.3882506191730499, "learning_rate": 8e-05, "loss": 1.7341, "step": 5236 }, { "epoch": 0.5659786015346374, "grad_norm": 0.48214930295944214, "learning_rate": 8e-05, "loss": 1.9205, "step": 5237 }, { "epoch": 0.5660866745920242, "grad_norm": 0.4377490282058716, "learning_rate": 8e-05, "loss": 1.5066, "step": 5238 }, { "epoch": 0.566194747649411, "grad_norm": 0.410521000623703, "learning_rate": 8e-05, "loss": 1.6053, "step": 5239 }, { "epoch": 0.5663028207067978, "grad_norm": 0.42951399087905884, "learning_rate": 8e-05, "loss": 1.8188, "step": 5240 }, { "epoch": 0.5664108937641846, "grad_norm": 0.41424447298049927, "learning_rate": 8e-05, "loss": 1.7034, "step": 5241 }, { "epoch": 0.5665189668215714, "grad_norm": 0.38773444294929504, "learning_rate": 8e-05, "loss": 1.5908, "step": 5242 }, { "epoch": 0.5666270398789581, "grad_norm": 0.40095382928848267, "learning_rate": 8e-05, "loss": 1.725, "step": 5243 }, { "epoch": 0.5667351129363449, "grad_norm": 0.3822839558124542, "learning_rate": 8e-05, "loss": 1.6679, "step": 5244 }, { "epoch": 0.5668431859937317, "grad_norm": 0.5377656817436218, "learning_rate": 8e-05, "loss": 1.9494, "step": 5245 }, { "epoch": 0.5669512590511185, "grad_norm": 0.44104647636413574, "learning_rate": 8e-05, "loss": 1.849, "step": 5246 }, { "epoch": 0.5670593321085053, "grad_norm": 0.4137105345726013, "learning_rate": 8e-05, "loss": 1.7547, "step": 5247 }, { "epoch": 0.5671674051658921, "grad_norm": 0.42235273122787476, "learning_rate": 8e-05, "loss": 1.8208, "step": 5248 }, { "epoch": 0.567275478223279, "grad_norm": 0.387228786945343, "learning_rate": 8e-05, "loss": 1.7641, "step": 5249 }, { "epoch": 0.5673835512806658, "grad_norm": 0.4441639184951782, "learning_rate": 8e-05, "loss": 1.8837, "step": 5250 }, { "epoch": 0.5674916243380526, "grad_norm": 0.38530638813972473, "learning_rate": 8e-05, "loss": 1.5735, "step": 5251 }, { "epoch": 0.5675996973954394, "grad_norm": 0.44352278113365173, "learning_rate": 8e-05, "loss": 1.7636, "step": 5252 }, { "epoch": 0.5677077704528262, "grad_norm": 0.43755489587783813, "learning_rate": 8e-05, "loss": 1.9073, "step": 5253 }, { "epoch": 0.5678158435102129, "grad_norm": 0.4125930070877075, "learning_rate": 8e-05, "loss": 1.6402, "step": 5254 }, { "epoch": 0.5679239165675997, "grad_norm": 0.42264389991760254, "learning_rate": 8e-05, "loss": 1.7613, "step": 5255 }, { "epoch": 0.5680319896249865, "grad_norm": 0.4340897500514984, "learning_rate": 8e-05, "loss": 1.8299, "step": 5256 }, { "epoch": 0.5681400626823733, "grad_norm": 0.6481916308403015, "learning_rate": 8e-05, "loss": 2.0455, "step": 5257 }, { "epoch": 0.5682481357397601, "grad_norm": 0.43799299001693726, "learning_rate": 8e-05, "loss": 1.7119, "step": 5258 }, { "epoch": 0.5683562087971469, "grad_norm": 0.4162825345993042, "learning_rate": 8e-05, "loss": 1.8124, "step": 5259 }, { "epoch": 0.5684642818545337, "grad_norm": 0.42771390080451965, "learning_rate": 8e-05, "loss": 1.8033, "step": 5260 }, { "epoch": 0.5685723549119205, "grad_norm": 0.38923880457878113, "learning_rate": 8e-05, "loss": 1.5863, "step": 5261 }, { "epoch": 0.5686804279693073, "grad_norm": 0.4318300187587738, "learning_rate": 8e-05, "loss": 1.6584, "step": 5262 }, { "epoch": 0.5687885010266941, "grad_norm": 0.39915651082992554, "learning_rate": 8e-05, "loss": 1.8064, "step": 5263 }, { "epoch": 0.5688965740840808, "grad_norm": 0.4172555208206177, "learning_rate": 8e-05, "loss": 1.5917, "step": 5264 }, { "epoch": 0.5690046471414676, "grad_norm": 0.4031795561313629, "learning_rate": 8e-05, "loss": 1.5238, "step": 5265 }, { "epoch": 0.5691127201988544, "grad_norm": 0.39053240418434143, "learning_rate": 8e-05, "loss": 1.5522, "step": 5266 }, { "epoch": 0.5692207932562412, "grad_norm": 0.43292075395584106, "learning_rate": 8e-05, "loss": 1.7589, "step": 5267 }, { "epoch": 0.569328866313628, "grad_norm": 0.39315733313560486, "learning_rate": 8e-05, "loss": 1.7806, "step": 5268 }, { "epoch": 0.5694369393710148, "grad_norm": 0.43249794840812683, "learning_rate": 8e-05, "loss": 1.6219, "step": 5269 }, { "epoch": 0.5695450124284016, "grad_norm": 0.3756706714630127, "learning_rate": 8e-05, "loss": 1.5041, "step": 5270 }, { "epoch": 0.5696530854857884, "grad_norm": 0.3935736417770386, "learning_rate": 8e-05, "loss": 1.6681, "step": 5271 }, { "epoch": 0.5697611585431752, "grad_norm": 0.4255791902542114, "learning_rate": 8e-05, "loss": 1.7855, "step": 5272 }, { "epoch": 0.569869231600562, "grad_norm": 0.40939441323280334, "learning_rate": 8e-05, "loss": 1.7338, "step": 5273 }, { "epoch": 0.5699773046579488, "grad_norm": 0.4053923487663269, "learning_rate": 8e-05, "loss": 1.5155, "step": 5274 }, { "epoch": 0.5700853777153355, "grad_norm": 0.3750765919685364, "learning_rate": 8e-05, "loss": 1.4862, "step": 5275 }, { "epoch": 0.5701934507727223, "grad_norm": 0.3933415114879608, "learning_rate": 8e-05, "loss": 1.7855, "step": 5276 }, { "epoch": 0.5703015238301091, "grad_norm": 0.4334651231765747, "learning_rate": 8e-05, "loss": 1.8461, "step": 5277 }, { "epoch": 0.5704095968874959, "grad_norm": 0.4651225507259369, "learning_rate": 8e-05, "loss": 1.8607, "step": 5278 }, { "epoch": 0.5705176699448827, "grad_norm": 0.4414561092853546, "learning_rate": 8e-05, "loss": 1.7453, "step": 5279 }, { "epoch": 0.5706257430022695, "grad_norm": 0.39202210307121277, "learning_rate": 8e-05, "loss": 1.6763, "step": 5280 }, { "epoch": 0.5707338160596563, "grad_norm": 0.40286022424697876, "learning_rate": 8e-05, "loss": 1.7098, "step": 5281 }, { "epoch": 0.5708418891170431, "grad_norm": 0.3997704088687897, "learning_rate": 8e-05, "loss": 1.6704, "step": 5282 }, { "epoch": 0.57094996217443, "grad_norm": 0.38595813512802124, "learning_rate": 8e-05, "loss": 1.6095, "step": 5283 }, { "epoch": 0.5710580352318168, "grad_norm": 0.4145796298980713, "learning_rate": 8e-05, "loss": 1.7508, "step": 5284 }, { "epoch": 0.5711661082892036, "grad_norm": 0.42384880781173706, "learning_rate": 8e-05, "loss": 1.7874, "step": 5285 }, { "epoch": 0.5712741813465902, "grad_norm": 0.3950011432170868, "learning_rate": 8e-05, "loss": 1.6387, "step": 5286 }, { "epoch": 0.571382254403977, "grad_norm": 0.3984391987323761, "learning_rate": 8e-05, "loss": 1.555, "step": 5287 }, { "epoch": 0.5714903274613639, "grad_norm": 0.38634058833122253, "learning_rate": 8e-05, "loss": 1.5851, "step": 5288 }, { "epoch": 0.5715984005187507, "grad_norm": 0.3852030634880066, "learning_rate": 8e-05, "loss": 1.6392, "step": 5289 }, { "epoch": 0.5717064735761375, "grad_norm": 0.4152202606201172, "learning_rate": 8e-05, "loss": 1.7308, "step": 5290 }, { "epoch": 0.5718145466335243, "grad_norm": 0.4210085868835449, "learning_rate": 8e-05, "loss": 1.674, "step": 5291 }, { "epoch": 0.5719226196909111, "grad_norm": 0.45221611857414246, "learning_rate": 8e-05, "loss": 1.8271, "step": 5292 }, { "epoch": 0.5720306927482979, "grad_norm": 0.4186042249202728, "learning_rate": 8e-05, "loss": 1.7451, "step": 5293 }, { "epoch": 0.5721387658056847, "grad_norm": 0.42520201206207275, "learning_rate": 8e-05, "loss": 1.7087, "step": 5294 }, { "epoch": 0.5722468388630715, "grad_norm": 0.4219418466091156, "learning_rate": 8e-05, "loss": 1.6405, "step": 5295 }, { "epoch": 0.5723549119204582, "grad_norm": 0.4172126054763794, "learning_rate": 8e-05, "loss": 1.7902, "step": 5296 }, { "epoch": 0.572462984977845, "grad_norm": 0.4092489778995514, "learning_rate": 8e-05, "loss": 1.7276, "step": 5297 }, { "epoch": 0.5725710580352318, "grad_norm": 0.40086889266967773, "learning_rate": 8e-05, "loss": 1.6391, "step": 5298 }, { "epoch": 0.5726791310926186, "grad_norm": 0.386207640171051, "learning_rate": 8e-05, "loss": 1.6958, "step": 5299 }, { "epoch": 0.5727872041500054, "grad_norm": 0.420908659696579, "learning_rate": 8e-05, "loss": 1.7793, "step": 5300 }, { "epoch": 0.5728952772073922, "grad_norm": 0.430510014295578, "learning_rate": 8e-05, "loss": 1.6023, "step": 5301 }, { "epoch": 0.573003350264779, "grad_norm": 0.41509032249450684, "learning_rate": 8e-05, "loss": 1.7761, "step": 5302 }, { "epoch": 0.5731114233221658, "grad_norm": 0.38490960001945496, "learning_rate": 8e-05, "loss": 1.6531, "step": 5303 }, { "epoch": 0.5732194963795526, "grad_norm": 0.41628769040107727, "learning_rate": 8e-05, "loss": 1.6441, "step": 5304 }, { "epoch": 0.5733275694369394, "grad_norm": 0.43089672923088074, "learning_rate": 8e-05, "loss": 1.5992, "step": 5305 }, { "epoch": 0.5734356424943262, "grad_norm": 0.4329648017883301, "learning_rate": 8e-05, "loss": 1.7796, "step": 5306 }, { "epoch": 0.5735437155517129, "grad_norm": 0.4258885383605957, "learning_rate": 8e-05, "loss": 1.8568, "step": 5307 }, { "epoch": 0.5736517886090997, "grad_norm": 0.42053696513175964, "learning_rate": 8e-05, "loss": 1.7144, "step": 5308 }, { "epoch": 0.5737598616664865, "grad_norm": 0.4371488690376282, "learning_rate": 8e-05, "loss": 1.8927, "step": 5309 }, { "epoch": 0.5738679347238733, "grad_norm": 0.4615253508090973, "learning_rate": 8e-05, "loss": 1.8143, "step": 5310 }, { "epoch": 0.5739760077812601, "grad_norm": 0.39801907539367676, "learning_rate": 8e-05, "loss": 1.6795, "step": 5311 }, { "epoch": 0.5740840808386469, "grad_norm": 0.4013351798057556, "learning_rate": 8e-05, "loss": 1.6555, "step": 5312 }, { "epoch": 0.5741921538960337, "grad_norm": 0.40145784616470337, "learning_rate": 8e-05, "loss": 1.6286, "step": 5313 }, { "epoch": 0.5743002269534205, "grad_norm": 0.4190628230571747, "learning_rate": 8e-05, "loss": 1.7657, "step": 5314 }, { "epoch": 0.5744083000108073, "grad_norm": 0.4427117705345154, "learning_rate": 8e-05, "loss": 1.6172, "step": 5315 }, { "epoch": 0.5745163730681941, "grad_norm": 0.3714866042137146, "learning_rate": 8e-05, "loss": 1.5914, "step": 5316 }, { "epoch": 0.574624446125581, "grad_norm": 0.44074395298957825, "learning_rate": 8e-05, "loss": 1.7816, "step": 5317 }, { "epoch": 0.5747325191829676, "grad_norm": 0.398269385099411, "learning_rate": 8e-05, "loss": 1.5183, "step": 5318 }, { "epoch": 0.5748405922403544, "grad_norm": 0.39543092250823975, "learning_rate": 8e-05, "loss": 1.6352, "step": 5319 }, { "epoch": 0.5749486652977412, "grad_norm": 0.385591059923172, "learning_rate": 8e-05, "loss": 1.5787, "step": 5320 }, { "epoch": 0.575056738355128, "grad_norm": 0.44630762934684753, "learning_rate": 8e-05, "loss": 1.7581, "step": 5321 }, { "epoch": 0.5751648114125149, "grad_norm": 0.44282087683677673, "learning_rate": 8e-05, "loss": 1.877, "step": 5322 }, { "epoch": 0.5752728844699017, "grad_norm": 0.42871883511543274, "learning_rate": 8e-05, "loss": 1.6977, "step": 5323 }, { "epoch": 0.5753809575272885, "grad_norm": 0.4049347937107086, "learning_rate": 8e-05, "loss": 1.5915, "step": 5324 }, { "epoch": 0.5754890305846753, "grad_norm": 0.438162237405777, "learning_rate": 8e-05, "loss": 1.8261, "step": 5325 }, { "epoch": 0.5755971036420621, "grad_norm": 0.4335150122642517, "learning_rate": 8e-05, "loss": 1.6951, "step": 5326 }, { "epoch": 0.5757051766994489, "grad_norm": 0.4028345048427582, "learning_rate": 8e-05, "loss": 1.629, "step": 5327 }, { "epoch": 0.5758132497568356, "grad_norm": 0.4086396396160126, "learning_rate": 8e-05, "loss": 1.7121, "step": 5328 }, { "epoch": 0.5759213228142224, "grad_norm": 0.4495050609111786, "learning_rate": 8e-05, "loss": 1.9981, "step": 5329 }, { "epoch": 0.5760293958716092, "grad_norm": 0.42505890130996704, "learning_rate": 8e-05, "loss": 1.8245, "step": 5330 }, { "epoch": 0.576137468928996, "grad_norm": 0.4961816966533661, "learning_rate": 8e-05, "loss": 1.629, "step": 5331 }, { "epoch": 0.5762455419863828, "grad_norm": 0.4229595363140106, "learning_rate": 8e-05, "loss": 1.728, "step": 5332 }, { "epoch": 0.5763536150437696, "grad_norm": 0.39661651849746704, "learning_rate": 8e-05, "loss": 1.5103, "step": 5333 }, { "epoch": 0.5764616881011564, "grad_norm": 0.4226754605770111, "learning_rate": 8e-05, "loss": 1.7556, "step": 5334 }, { "epoch": 0.5765697611585432, "grad_norm": 0.4355357587337494, "learning_rate": 8e-05, "loss": 1.6101, "step": 5335 }, { "epoch": 0.57667783421593, "grad_norm": 0.39232200384140015, "learning_rate": 8e-05, "loss": 1.7082, "step": 5336 }, { "epoch": 0.5767859072733168, "grad_norm": 0.44198891520500183, "learning_rate": 8e-05, "loss": 1.648, "step": 5337 }, { "epoch": 0.5768939803307036, "grad_norm": 0.4066906273365021, "learning_rate": 8e-05, "loss": 1.577, "step": 5338 }, { "epoch": 0.5770020533880903, "grad_norm": 0.3861295282840729, "learning_rate": 8e-05, "loss": 1.6432, "step": 5339 }, { "epoch": 0.5771101264454771, "grad_norm": 0.3998039960861206, "learning_rate": 8e-05, "loss": 1.5501, "step": 5340 }, { "epoch": 0.5772181995028639, "grad_norm": 0.4161585569381714, "learning_rate": 8e-05, "loss": 1.609, "step": 5341 }, { "epoch": 0.5773262725602507, "grad_norm": 0.4201744794845581, "learning_rate": 8e-05, "loss": 1.7017, "step": 5342 }, { "epoch": 0.5774343456176375, "grad_norm": 0.3946765661239624, "learning_rate": 8e-05, "loss": 1.6548, "step": 5343 }, { "epoch": 0.5775424186750243, "grad_norm": 0.40827125310897827, "learning_rate": 8e-05, "loss": 1.6112, "step": 5344 }, { "epoch": 0.5776504917324111, "grad_norm": 0.4174119532108307, "learning_rate": 8e-05, "loss": 1.7782, "step": 5345 }, { "epoch": 0.5777585647897979, "grad_norm": 0.4209985136985779, "learning_rate": 8e-05, "loss": 1.7371, "step": 5346 }, { "epoch": 0.5778666378471847, "grad_norm": 0.4366307258605957, "learning_rate": 8e-05, "loss": 1.7216, "step": 5347 }, { "epoch": 0.5779747109045715, "grad_norm": 0.4198886752128601, "learning_rate": 8e-05, "loss": 1.7205, "step": 5348 }, { "epoch": 0.5780827839619583, "grad_norm": 0.4105747938156128, "learning_rate": 8e-05, "loss": 1.7029, "step": 5349 }, { "epoch": 0.578190857019345, "grad_norm": 0.41520342230796814, "learning_rate": 8e-05, "loss": 1.6988, "step": 5350 }, { "epoch": 0.5782989300767318, "grad_norm": 0.3858321011066437, "learning_rate": 8e-05, "loss": 1.5534, "step": 5351 }, { "epoch": 0.5784070031341186, "grad_norm": 0.3844169080257416, "learning_rate": 8e-05, "loss": 1.4212, "step": 5352 }, { "epoch": 0.5785150761915054, "grad_norm": 0.449170857667923, "learning_rate": 8e-05, "loss": 1.7656, "step": 5353 }, { "epoch": 0.5786231492488922, "grad_norm": 0.4232487678527832, "learning_rate": 8e-05, "loss": 1.5628, "step": 5354 }, { "epoch": 0.578731222306279, "grad_norm": 0.41104230284690857, "learning_rate": 8e-05, "loss": 1.7439, "step": 5355 }, { "epoch": 0.5788392953636659, "grad_norm": 0.3864450454711914, "learning_rate": 8e-05, "loss": 1.6222, "step": 5356 }, { "epoch": 0.5789473684210527, "grad_norm": 0.4142056405544281, "learning_rate": 8e-05, "loss": 1.7881, "step": 5357 }, { "epoch": 0.5790554414784395, "grad_norm": 0.397704541683197, "learning_rate": 8e-05, "loss": 1.5933, "step": 5358 }, { "epoch": 0.5791635145358263, "grad_norm": 0.3921453654766083, "learning_rate": 8e-05, "loss": 1.6443, "step": 5359 }, { "epoch": 0.5792715875932131, "grad_norm": 0.40960490703582764, "learning_rate": 8e-05, "loss": 1.6124, "step": 5360 }, { "epoch": 0.5793796606505998, "grad_norm": 0.436685711145401, "learning_rate": 8e-05, "loss": 1.6312, "step": 5361 }, { "epoch": 0.5794877337079866, "grad_norm": 0.40005064010620117, "learning_rate": 8e-05, "loss": 1.7092, "step": 5362 }, { "epoch": 0.5795958067653734, "grad_norm": 0.4011305570602417, "learning_rate": 8e-05, "loss": 1.5461, "step": 5363 }, { "epoch": 0.5797038798227602, "grad_norm": 0.4219714403152466, "learning_rate": 8e-05, "loss": 1.8157, "step": 5364 }, { "epoch": 0.579811952880147, "grad_norm": 0.43419039249420166, "learning_rate": 8e-05, "loss": 1.7713, "step": 5365 }, { "epoch": 0.5799200259375338, "grad_norm": 0.4255037009716034, "learning_rate": 8e-05, "loss": 1.5958, "step": 5366 }, { "epoch": 0.5800280989949206, "grad_norm": 0.4723442792892456, "learning_rate": 8e-05, "loss": 2.0607, "step": 5367 }, { "epoch": 0.5801361720523074, "grad_norm": 0.40883538126945496, "learning_rate": 8e-05, "loss": 1.571, "step": 5368 }, { "epoch": 0.5802442451096942, "grad_norm": 0.42653220891952515, "learning_rate": 8e-05, "loss": 1.8421, "step": 5369 }, { "epoch": 0.580352318167081, "grad_norm": 0.4188445210456848, "learning_rate": 8e-05, "loss": 1.4766, "step": 5370 }, { "epoch": 0.5804603912244677, "grad_norm": 0.43071654438972473, "learning_rate": 8e-05, "loss": 1.7116, "step": 5371 }, { "epoch": 0.5805684642818545, "grad_norm": 0.3859540820121765, "learning_rate": 8e-05, "loss": 1.5893, "step": 5372 }, { "epoch": 0.5806765373392413, "grad_norm": 0.36934009194374084, "learning_rate": 8e-05, "loss": 1.4969, "step": 5373 }, { "epoch": 0.5807846103966281, "grad_norm": 0.41599804162979126, "learning_rate": 8e-05, "loss": 1.7266, "step": 5374 }, { "epoch": 0.5808926834540149, "grad_norm": 0.4604172706604004, "learning_rate": 8e-05, "loss": 1.835, "step": 5375 }, { "epoch": 0.5810007565114017, "grad_norm": 0.456888884305954, "learning_rate": 8e-05, "loss": 1.6002, "step": 5376 }, { "epoch": 0.5811088295687885, "grad_norm": 0.40447938442230225, "learning_rate": 8e-05, "loss": 1.6817, "step": 5377 }, { "epoch": 0.5812169026261753, "grad_norm": 0.4057120680809021, "learning_rate": 8e-05, "loss": 1.6781, "step": 5378 }, { "epoch": 0.5813249756835621, "grad_norm": 0.4018944799900055, "learning_rate": 8e-05, "loss": 1.6266, "step": 5379 }, { "epoch": 0.5814330487409489, "grad_norm": 0.40940529108047485, "learning_rate": 8e-05, "loss": 1.7143, "step": 5380 }, { "epoch": 0.5815411217983357, "grad_norm": 0.39011016488075256, "learning_rate": 8e-05, "loss": 1.6104, "step": 5381 }, { "epoch": 0.5816491948557224, "grad_norm": 0.3987281024456024, "learning_rate": 8e-05, "loss": 1.6036, "step": 5382 }, { "epoch": 0.5817572679131092, "grad_norm": 0.43238675594329834, "learning_rate": 8e-05, "loss": 1.7365, "step": 5383 }, { "epoch": 0.581865340970496, "grad_norm": 0.3840087354183197, "learning_rate": 8e-05, "loss": 1.6776, "step": 5384 }, { "epoch": 0.5819734140278828, "grad_norm": 0.3946155905723572, "learning_rate": 8e-05, "loss": 1.5803, "step": 5385 }, { "epoch": 0.5820814870852696, "grad_norm": 0.4202658236026764, "learning_rate": 8e-05, "loss": 1.6876, "step": 5386 }, { "epoch": 0.5821895601426564, "grad_norm": 0.37610161304473877, "learning_rate": 8e-05, "loss": 1.5442, "step": 5387 }, { "epoch": 0.5822976332000432, "grad_norm": 0.39067313075065613, "learning_rate": 8e-05, "loss": 1.6252, "step": 5388 }, { "epoch": 0.58240570625743, "grad_norm": 0.45235398411750793, "learning_rate": 8e-05, "loss": 1.8751, "step": 5389 }, { "epoch": 0.5825137793148168, "grad_norm": 0.37975701689720154, "learning_rate": 8e-05, "loss": 1.481, "step": 5390 }, { "epoch": 0.5826218523722037, "grad_norm": 0.4280672073364258, "learning_rate": 8e-05, "loss": 1.5631, "step": 5391 }, { "epoch": 0.5827299254295905, "grad_norm": 0.384395033121109, "learning_rate": 8e-05, "loss": 1.5212, "step": 5392 }, { "epoch": 0.5828379984869771, "grad_norm": 0.38932591676712036, "learning_rate": 8e-05, "loss": 1.5088, "step": 5393 }, { "epoch": 0.582946071544364, "grad_norm": 0.44221773743629456, "learning_rate": 8e-05, "loss": 1.6158, "step": 5394 }, { "epoch": 0.5830541446017508, "grad_norm": 0.394045889377594, "learning_rate": 8e-05, "loss": 1.5628, "step": 5395 }, { "epoch": 0.5831622176591376, "grad_norm": 0.42171511054039, "learning_rate": 8e-05, "loss": 1.657, "step": 5396 }, { "epoch": 0.5832702907165244, "grad_norm": 0.4027475416660309, "learning_rate": 8e-05, "loss": 1.614, "step": 5397 }, { "epoch": 0.5833783637739112, "grad_norm": 0.44133836030960083, "learning_rate": 8e-05, "loss": 1.6019, "step": 5398 }, { "epoch": 0.583486436831298, "grad_norm": 0.4574414789676666, "learning_rate": 8e-05, "loss": 1.6581, "step": 5399 }, { "epoch": 0.5835945098886848, "grad_norm": 0.39825439453125, "learning_rate": 8e-05, "loss": 1.6174, "step": 5400 }, { "epoch": 0.5837025829460716, "grad_norm": 0.4111921787261963, "learning_rate": 8e-05, "loss": 1.6539, "step": 5401 }, { "epoch": 0.5838106560034584, "grad_norm": 0.44841358065605164, "learning_rate": 8e-05, "loss": 1.6875, "step": 5402 }, { "epoch": 0.5839187290608451, "grad_norm": 0.43147170543670654, "learning_rate": 8e-05, "loss": 1.6804, "step": 5403 }, { "epoch": 0.5840268021182319, "grad_norm": 0.4126768410205841, "learning_rate": 8e-05, "loss": 1.7614, "step": 5404 }, { "epoch": 0.5841348751756187, "grad_norm": 0.49587559700012207, "learning_rate": 8e-05, "loss": 1.9461, "step": 5405 }, { "epoch": 0.5842429482330055, "grad_norm": 0.41558846831321716, "learning_rate": 8e-05, "loss": 1.696, "step": 5406 }, { "epoch": 0.5843510212903923, "grad_norm": 0.42491477727890015, "learning_rate": 8e-05, "loss": 1.749, "step": 5407 }, { "epoch": 0.5844590943477791, "grad_norm": 0.44294315576553345, "learning_rate": 8e-05, "loss": 1.6565, "step": 5408 }, { "epoch": 0.5845671674051659, "grad_norm": 0.4336879849433899, "learning_rate": 8e-05, "loss": 1.7422, "step": 5409 }, { "epoch": 0.5846752404625527, "grad_norm": 0.3901079297065735, "learning_rate": 8e-05, "loss": 1.6658, "step": 5410 }, { "epoch": 0.5847833135199395, "grad_norm": 0.3946743607521057, "learning_rate": 8e-05, "loss": 1.6099, "step": 5411 }, { "epoch": 0.5848913865773263, "grad_norm": 0.42238330841064453, "learning_rate": 8e-05, "loss": 1.7042, "step": 5412 }, { "epoch": 0.5849994596347131, "grad_norm": 0.4060431122779846, "learning_rate": 8e-05, "loss": 1.8187, "step": 5413 }, { "epoch": 0.5851075326920998, "grad_norm": 0.38780081272125244, "learning_rate": 8e-05, "loss": 1.588, "step": 5414 }, { "epoch": 0.5852156057494866, "grad_norm": 0.4461301565170288, "learning_rate": 8e-05, "loss": 1.6291, "step": 5415 }, { "epoch": 0.5853236788068734, "grad_norm": 0.5144010186195374, "learning_rate": 8e-05, "loss": 1.9924, "step": 5416 }, { "epoch": 0.5854317518642602, "grad_norm": 0.4668757915496826, "learning_rate": 8e-05, "loss": 1.9669, "step": 5417 }, { "epoch": 0.585539824921647, "grad_norm": 0.40000119805336, "learning_rate": 8e-05, "loss": 1.6961, "step": 5418 }, { "epoch": 0.5856478979790338, "grad_norm": 0.42492106556892395, "learning_rate": 8e-05, "loss": 1.6386, "step": 5419 }, { "epoch": 0.5857559710364206, "grad_norm": 0.44114789366722107, "learning_rate": 8e-05, "loss": 1.7595, "step": 5420 }, { "epoch": 0.5858640440938074, "grad_norm": 0.42115095257759094, "learning_rate": 8e-05, "loss": 1.7825, "step": 5421 }, { "epoch": 0.5859721171511942, "grad_norm": 0.4470484256744385, "learning_rate": 8e-05, "loss": 1.8712, "step": 5422 }, { "epoch": 0.586080190208581, "grad_norm": 0.43827539682388306, "learning_rate": 8e-05, "loss": 1.7889, "step": 5423 }, { "epoch": 0.5861882632659678, "grad_norm": 0.4696293771266937, "learning_rate": 8e-05, "loss": 1.7043, "step": 5424 }, { "epoch": 0.5862963363233545, "grad_norm": 0.41933178901672363, "learning_rate": 8e-05, "loss": 1.6624, "step": 5425 }, { "epoch": 0.5864044093807413, "grad_norm": 0.4222598373889923, "learning_rate": 8e-05, "loss": 1.7518, "step": 5426 }, { "epoch": 0.5865124824381281, "grad_norm": 0.4163535237312317, "learning_rate": 8e-05, "loss": 1.6871, "step": 5427 }, { "epoch": 0.586620555495515, "grad_norm": 0.42308667302131653, "learning_rate": 8e-05, "loss": 1.6836, "step": 5428 }, { "epoch": 0.5867286285529018, "grad_norm": 0.4433268904685974, "learning_rate": 8e-05, "loss": 1.7637, "step": 5429 }, { "epoch": 0.5868367016102886, "grad_norm": 0.4544048011302948, "learning_rate": 8e-05, "loss": 1.7523, "step": 5430 }, { "epoch": 0.5869447746676754, "grad_norm": 0.46462416648864746, "learning_rate": 8e-05, "loss": 1.7743, "step": 5431 }, { "epoch": 0.5870528477250622, "grad_norm": 0.40328723192214966, "learning_rate": 8e-05, "loss": 1.7062, "step": 5432 }, { "epoch": 0.587160920782449, "grad_norm": 0.4461677372455597, "learning_rate": 8e-05, "loss": 1.768, "step": 5433 }, { "epoch": 0.5872689938398358, "grad_norm": 0.4004385769367218, "learning_rate": 8e-05, "loss": 1.5617, "step": 5434 }, { "epoch": 0.5873770668972226, "grad_norm": 0.429597944021225, "learning_rate": 8e-05, "loss": 1.7134, "step": 5435 }, { "epoch": 0.5874851399546093, "grad_norm": 0.41334402561187744, "learning_rate": 8e-05, "loss": 1.7523, "step": 5436 }, { "epoch": 0.5875932130119961, "grad_norm": 0.41952815651893616, "learning_rate": 8e-05, "loss": 1.7976, "step": 5437 }, { "epoch": 0.5877012860693829, "grad_norm": 0.42964908480644226, "learning_rate": 8e-05, "loss": 1.8072, "step": 5438 }, { "epoch": 0.5878093591267697, "grad_norm": 0.4672430753707886, "learning_rate": 8e-05, "loss": 1.8278, "step": 5439 }, { "epoch": 0.5879174321841565, "grad_norm": 0.42697086930274963, "learning_rate": 8e-05, "loss": 1.7137, "step": 5440 }, { "epoch": 0.5880255052415433, "grad_norm": 0.4287816882133484, "learning_rate": 8e-05, "loss": 1.8057, "step": 5441 }, { "epoch": 0.5881335782989301, "grad_norm": 0.48263218998908997, "learning_rate": 8e-05, "loss": 1.8319, "step": 5442 }, { "epoch": 0.5882416513563169, "grad_norm": 0.40174001455307007, "learning_rate": 8e-05, "loss": 1.6672, "step": 5443 }, { "epoch": 0.5883497244137037, "grad_norm": 0.3892255425453186, "learning_rate": 8e-05, "loss": 1.6032, "step": 5444 }, { "epoch": 0.5884577974710905, "grad_norm": 0.4220700263977051, "learning_rate": 8e-05, "loss": 1.6116, "step": 5445 }, { "epoch": 0.5885658705284772, "grad_norm": 0.3978031575679779, "learning_rate": 8e-05, "loss": 1.5733, "step": 5446 }, { "epoch": 0.588673943585864, "grad_norm": 0.448760449886322, "learning_rate": 8e-05, "loss": 1.7832, "step": 5447 }, { "epoch": 0.5887820166432508, "grad_norm": 0.3821183145046234, "learning_rate": 8e-05, "loss": 1.4467, "step": 5448 }, { "epoch": 0.5888900897006376, "grad_norm": 0.411353200674057, "learning_rate": 8e-05, "loss": 1.4766, "step": 5449 }, { "epoch": 0.5889981627580244, "grad_norm": 0.3983737826347351, "learning_rate": 8e-05, "loss": 1.6907, "step": 5450 }, { "epoch": 0.5891062358154112, "grad_norm": 0.41489923000335693, "learning_rate": 8e-05, "loss": 1.7962, "step": 5451 }, { "epoch": 0.589214308872798, "grad_norm": 0.41959306597709656, "learning_rate": 8e-05, "loss": 1.7185, "step": 5452 }, { "epoch": 0.5893223819301848, "grad_norm": 0.43487170338630676, "learning_rate": 8e-05, "loss": 1.6241, "step": 5453 }, { "epoch": 0.5894304549875716, "grad_norm": 0.3847074508666992, "learning_rate": 8e-05, "loss": 1.5842, "step": 5454 }, { "epoch": 0.5895385280449584, "grad_norm": 0.43493595719337463, "learning_rate": 8e-05, "loss": 1.7194, "step": 5455 }, { "epoch": 0.5896466011023452, "grad_norm": 0.4085455536842346, "learning_rate": 8e-05, "loss": 1.7903, "step": 5456 }, { "epoch": 0.5897546741597319, "grad_norm": 0.4117777347564697, "learning_rate": 8e-05, "loss": 1.6498, "step": 5457 }, { "epoch": 0.5898627472171187, "grad_norm": 0.44066834449768066, "learning_rate": 8e-05, "loss": 1.8277, "step": 5458 }, { "epoch": 0.5899708202745055, "grad_norm": 0.4604656994342804, "learning_rate": 8e-05, "loss": 1.784, "step": 5459 }, { "epoch": 0.5900788933318923, "grad_norm": 0.45529627799987793, "learning_rate": 8e-05, "loss": 1.6549, "step": 5460 }, { "epoch": 0.5901869663892791, "grad_norm": 0.441607266664505, "learning_rate": 8e-05, "loss": 1.7379, "step": 5461 }, { "epoch": 0.590295039446666, "grad_norm": 0.4331668019294739, "learning_rate": 8e-05, "loss": 1.6217, "step": 5462 }, { "epoch": 0.5904031125040528, "grad_norm": 0.4128395915031433, "learning_rate": 8e-05, "loss": 1.6796, "step": 5463 }, { "epoch": 0.5905111855614396, "grad_norm": 0.48351430892944336, "learning_rate": 8e-05, "loss": 1.8196, "step": 5464 }, { "epoch": 0.5906192586188264, "grad_norm": 0.4197305738925934, "learning_rate": 8e-05, "loss": 1.6693, "step": 5465 }, { "epoch": 0.5907273316762132, "grad_norm": 0.37849923968315125, "learning_rate": 8e-05, "loss": 1.5463, "step": 5466 }, { "epoch": 0.5908354047336, "grad_norm": 0.41002747416496277, "learning_rate": 8e-05, "loss": 1.6505, "step": 5467 }, { "epoch": 0.5909434777909867, "grad_norm": 0.43111175298690796, "learning_rate": 8e-05, "loss": 1.826, "step": 5468 }, { "epoch": 0.5910515508483735, "grad_norm": 0.4639660716056824, "learning_rate": 8e-05, "loss": 1.7833, "step": 5469 }, { "epoch": 0.5911596239057603, "grad_norm": 0.43614310026168823, "learning_rate": 8e-05, "loss": 1.7349, "step": 5470 }, { "epoch": 0.5912676969631471, "grad_norm": 0.4726816713809967, "learning_rate": 8e-05, "loss": 1.8066, "step": 5471 }, { "epoch": 0.5913757700205339, "grad_norm": 0.4578784108161926, "learning_rate": 8e-05, "loss": 1.8504, "step": 5472 }, { "epoch": 0.5914838430779207, "grad_norm": 0.39253348112106323, "learning_rate": 8e-05, "loss": 1.6897, "step": 5473 }, { "epoch": 0.5915919161353075, "grad_norm": 0.4816431999206543, "learning_rate": 8e-05, "loss": 1.764, "step": 5474 }, { "epoch": 0.5916999891926943, "grad_norm": 0.3907853066921234, "learning_rate": 8e-05, "loss": 1.5757, "step": 5475 }, { "epoch": 0.5918080622500811, "grad_norm": 0.4162692129611969, "learning_rate": 8e-05, "loss": 1.5515, "step": 5476 }, { "epoch": 0.5919161353074679, "grad_norm": 0.41307544708251953, "learning_rate": 8e-05, "loss": 1.5795, "step": 5477 }, { "epoch": 0.5920242083648546, "grad_norm": 0.4000306725502014, "learning_rate": 8e-05, "loss": 1.7065, "step": 5478 }, { "epoch": 0.5921322814222414, "grad_norm": 0.478236585855484, "learning_rate": 8e-05, "loss": 1.8607, "step": 5479 }, { "epoch": 0.5922403544796282, "grad_norm": 0.42933961749076843, "learning_rate": 8e-05, "loss": 1.6035, "step": 5480 }, { "epoch": 0.592348427537015, "grad_norm": 0.41460591554641724, "learning_rate": 8e-05, "loss": 1.7603, "step": 5481 }, { "epoch": 0.5924565005944018, "grad_norm": 0.3942809998989105, "learning_rate": 8e-05, "loss": 1.599, "step": 5482 }, { "epoch": 0.5925645736517886, "grad_norm": 0.41784733533859253, "learning_rate": 8e-05, "loss": 1.6919, "step": 5483 }, { "epoch": 0.5926726467091754, "grad_norm": 0.4404323697090149, "learning_rate": 8e-05, "loss": 1.7307, "step": 5484 }, { "epoch": 0.5927807197665622, "grad_norm": 0.4154611825942993, "learning_rate": 8e-05, "loss": 1.7309, "step": 5485 }, { "epoch": 0.592888792823949, "grad_norm": 0.4442439079284668, "learning_rate": 8e-05, "loss": 1.8035, "step": 5486 }, { "epoch": 0.5929968658813358, "grad_norm": 0.43773746490478516, "learning_rate": 8e-05, "loss": 1.7771, "step": 5487 }, { "epoch": 0.5931049389387226, "grad_norm": 0.41343194246292114, "learning_rate": 8e-05, "loss": 1.7041, "step": 5488 }, { "epoch": 0.5932130119961093, "grad_norm": 0.4463178217411041, "learning_rate": 8e-05, "loss": 1.8569, "step": 5489 }, { "epoch": 0.5933210850534961, "grad_norm": 0.43044400215148926, "learning_rate": 8e-05, "loss": 1.9069, "step": 5490 }, { "epoch": 0.5934291581108829, "grad_norm": 0.3882802426815033, "learning_rate": 8e-05, "loss": 1.6138, "step": 5491 }, { "epoch": 0.5935372311682697, "grad_norm": 0.40134260058403015, "learning_rate": 8e-05, "loss": 1.7681, "step": 5492 }, { "epoch": 0.5936453042256565, "grad_norm": 0.41762566566467285, "learning_rate": 8e-05, "loss": 1.6939, "step": 5493 }, { "epoch": 0.5937533772830433, "grad_norm": 0.393058717250824, "learning_rate": 8e-05, "loss": 1.5012, "step": 5494 }, { "epoch": 0.5938614503404301, "grad_norm": 0.3982808291912079, "learning_rate": 8e-05, "loss": 1.6581, "step": 5495 }, { "epoch": 0.593969523397817, "grad_norm": 0.41006502509117126, "learning_rate": 8e-05, "loss": 1.7844, "step": 5496 }, { "epoch": 0.5940775964552037, "grad_norm": 0.42726224660873413, "learning_rate": 8e-05, "loss": 1.7747, "step": 5497 }, { "epoch": 0.5941856695125906, "grad_norm": 0.4124722480773926, "learning_rate": 8e-05, "loss": 1.6264, "step": 5498 }, { "epoch": 0.5942937425699774, "grad_norm": 0.40958869457244873, "learning_rate": 8e-05, "loss": 1.6071, "step": 5499 }, { "epoch": 0.594401815627364, "grad_norm": 0.40667110681533813, "learning_rate": 8e-05, "loss": 1.7236, "step": 5500 }, { "epoch": 0.5945098886847509, "grad_norm": 0.4166122078895569, "learning_rate": 8e-05, "loss": 1.6283, "step": 5501 }, { "epoch": 0.5946179617421377, "grad_norm": 0.457497775554657, "learning_rate": 8e-05, "loss": 1.7683, "step": 5502 }, { "epoch": 0.5947260347995245, "grad_norm": 0.44114089012145996, "learning_rate": 8e-05, "loss": 1.7253, "step": 5503 }, { "epoch": 0.5948341078569113, "grad_norm": 0.4012099504470825, "learning_rate": 8e-05, "loss": 1.5387, "step": 5504 }, { "epoch": 0.5949421809142981, "grad_norm": 0.4515479803085327, "learning_rate": 8e-05, "loss": 1.6906, "step": 5505 }, { "epoch": 0.5950502539716849, "grad_norm": 0.4611688256263733, "learning_rate": 8e-05, "loss": 1.8576, "step": 5506 }, { "epoch": 0.5951583270290717, "grad_norm": 0.42567819356918335, "learning_rate": 8e-05, "loss": 1.8528, "step": 5507 }, { "epoch": 0.5952664000864585, "grad_norm": 0.4588940739631653, "learning_rate": 8e-05, "loss": 1.9017, "step": 5508 }, { "epoch": 0.5953744731438453, "grad_norm": 0.41879788041114807, "learning_rate": 8e-05, "loss": 1.6013, "step": 5509 }, { "epoch": 0.5954825462012321, "grad_norm": 0.4030558466911316, "learning_rate": 8e-05, "loss": 1.6915, "step": 5510 }, { "epoch": 0.5955906192586188, "grad_norm": 0.39725369215011597, "learning_rate": 8e-05, "loss": 1.6942, "step": 5511 }, { "epoch": 0.5956986923160056, "grad_norm": 0.4687209129333496, "learning_rate": 8e-05, "loss": 1.9083, "step": 5512 }, { "epoch": 0.5958067653733924, "grad_norm": 0.4348128139972687, "learning_rate": 8e-05, "loss": 1.8668, "step": 5513 }, { "epoch": 0.5959148384307792, "grad_norm": 0.4291481375694275, "learning_rate": 8e-05, "loss": 1.6146, "step": 5514 }, { "epoch": 0.596022911488166, "grad_norm": 0.41463911533355713, "learning_rate": 8e-05, "loss": 1.6631, "step": 5515 }, { "epoch": 0.5961309845455528, "grad_norm": 0.426289826631546, "learning_rate": 8e-05, "loss": 1.8298, "step": 5516 }, { "epoch": 0.5962390576029396, "grad_norm": 0.4209468960762024, "learning_rate": 8e-05, "loss": 1.8054, "step": 5517 }, { "epoch": 0.5963471306603264, "grad_norm": 0.40028631687164307, "learning_rate": 8e-05, "loss": 1.7262, "step": 5518 }, { "epoch": 0.5964552037177132, "grad_norm": 0.4007244408130646, "learning_rate": 8e-05, "loss": 1.6435, "step": 5519 }, { "epoch": 0.5965632767751, "grad_norm": 0.45070523023605347, "learning_rate": 8e-05, "loss": 1.8989, "step": 5520 }, { "epoch": 0.5966713498324867, "grad_norm": 0.40943849086761475, "learning_rate": 8e-05, "loss": 1.6828, "step": 5521 }, { "epoch": 0.5967794228898735, "grad_norm": 0.4144580662250519, "learning_rate": 8e-05, "loss": 1.7319, "step": 5522 }, { "epoch": 0.5968874959472603, "grad_norm": 0.41063132882118225, "learning_rate": 8e-05, "loss": 1.817, "step": 5523 }, { "epoch": 0.5969955690046471, "grad_norm": 0.39429259300231934, "learning_rate": 8e-05, "loss": 1.51, "step": 5524 }, { "epoch": 0.5971036420620339, "grad_norm": 0.4246419668197632, "learning_rate": 8e-05, "loss": 1.6549, "step": 5525 }, { "epoch": 0.5972117151194207, "grad_norm": 0.4183419942855835, "learning_rate": 8e-05, "loss": 1.6206, "step": 5526 }, { "epoch": 0.5973197881768075, "grad_norm": 0.41949984431266785, "learning_rate": 8e-05, "loss": 1.7144, "step": 5527 }, { "epoch": 0.5974278612341943, "grad_norm": 0.43977347016334534, "learning_rate": 8e-05, "loss": 1.755, "step": 5528 }, { "epoch": 0.5975359342915811, "grad_norm": 0.4253687858581543, "learning_rate": 8e-05, "loss": 1.602, "step": 5529 }, { "epoch": 0.5976440073489679, "grad_norm": 0.4172111749649048, "learning_rate": 8e-05, "loss": 1.7304, "step": 5530 }, { "epoch": 0.5977520804063547, "grad_norm": 0.47139066457748413, "learning_rate": 8e-05, "loss": 1.8481, "step": 5531 }, { "epoch": 0.5978601534637414, "grad_norm": 0.4179459810256958, "learning_rate": 8e-05, "loss": 1.7879, "step": 5532 }, { "epoch": 0.5979682265211282, "grad_norm": 0.4005166292190552, "learning_rate": 8e-05, "loss": 1.7416, "step": 5533 }, { "epoch": 0.598076299578515, "grad_norm": 0.4050544202327728, "learning_rate": 8e-05, "loss": 1.8353, "step": 5534 }, { "epoch": 0.5981843726359019, "grad_norm": 0.4083227813243866, "learning_rate": 8e-05, "loss": 1.7889, "step": 5535 }, { "epoch": 0.5982924456932887, "grad_norm": 0.4093951880931854, "learning_rate": 8e-05, "loss": 1.8161, "step": 5536 }, { "epoch": 0.5984005187506755, "grad_norm": 0.38566264510154724, "learning_rate": 8e-05, "loss": 1.6979, "step": 5537 }, { "epoch": 0.5985085918080623, "grad_norm": 0.4569816291332245, "learning_rate": 8e-05, "loss": 1.8285, "step": 5538 }, { "epoch": 0.5986166648654491, "grad_norm": 0.4492870271205902, "learning_rate": 8e-05, "loss": 1.7771, "step": 5539 }, { "epoch": 0.5987247379228359, "grad_norm": 0.5212683081626892, "learning_rate": 8e-05, "loss": 1.6218, "step": 5540 }, { "epoch": 0.5988328109802227, "grad_norm": 0.44963589310646057, "learning_rate": 8e-05, "loss": 1.814, "step": 5541 }, { "epoch": 0.5989408840376095, "grad_norm": 0.4011872112751007, "learning_rate": 8e-05, "loss": 1.7146, "step": 5542 }, { "epoch": 0.5990489570949962, "grad_norm": 0.4305901527404785, "learning_rate": 8e-05, "loss": 1.6282, "step": 5543 }, { "epoch": 0.599157030152383, "grad_norm": 0.4531443119049072, "learning_rate": 8e-05, "loss": 1.8198, "step": 5544 }, { "epoch": 0.5992651032097698, "grad_norm": 0.39423680305480957, "learning_rate": 8e-05, "loss": 1.6705, "step": 5545 }, { "epoch": 0.5993731762671566, "grad_norm": 0.4384159445762634, "learning_rate": 8e-05, "loss": 1.7944, "step": 5546 }, { "epoch": 0.5994812493245434, "grad_norm": 0.49054139852523804, "learning_rate": 8e-05, "loss": 1.9136, "step": 5547 }, { "epoch": 0.5995893223819302, "grad_norm": 0.3991737961769104, "learning_rate": 8e-05, "loss": 1.6436, "step": 5548 }, { "epoch": 0.599697395439317, "grad_norm": 0.4394117295742035, "learning_rate": 8e-05, "loss": 1.7208, "step": 5549 }, { "epoch": 0.5998054684967038, "grad_norm": 0.38608548045158386, "learning_rate": 8e-05, "loss": 1.6287, "step": 5550 }, { "epoch": 0.5999135415540906, "grad_norm": 0.4394821226596832, "learning_rate": 8e-05, "loss": 1.7149, "step": 5551 }, { "epoch": 0.6000216146114774, "grad_norm": 0.40440675616264343, "learning_rate": 8e-05, "loss": 1.7971, "step": 5552 }, { "epoch": 0.6001296876688641, "grad_norm": 0.4036799967288971, "learning_rate": 8e-05, "loss": 1.5135, "step": 5553 }, { "epoch": 0.6002377607262509, "grad_norm": 0.40975648164749146, "learning_rate": 8e-05, "loss": 1.7428, "step": 5554 }, { "epoch": 0.6003458337836377, "grad_norm": 0.4443605840206146, "learning_rate": 8e-05, "loss": 1.7705, "step": 5555 }, { "epoch": 0.6004539068410245, "grad_norm": 0.4662593901157379, "learning_rate": 8e-05, "loss": 1.8267, "step": 5556 }, { "epoch": 0.6005619798984113, "grad_norm": 0.4011504352092743, "learning_rate": 8e-05, "loss": 1.5854, "step": 5557 }, { "epoch": 0.6006700529557981, "grad_norm": 0.3972177505493164, "learning_rate": 8e-05, "loss": 1.7554, "step": 5558 }, { "epoch": 0.6007781260131849, "grad_norm": 0.421519011259079, "learning_rate": 8e-05, "loss": 1.5552, "step": 5559 }, { "epoch": 0.6008861990705717, "grad_norm": 0.4157339930534363, "learning_rate": 8e-05, "loss": 1.6839, "step": 5560 }, { "epoch": 0.6009942721279585, "grad_norm": 0.4171294569969177, "learning_rate": 8e-05, "loss": 1.6965, "step": 5561 }, { "epoch": 0.6011023451853453, "grad_norm": 0.3961169123649597, "learning_rate": 8e-05, "loss": 1.5465, "step": 5562 }, { "epoch": 0.6012104182427321, "grad_norm": 0.43461012840270996, "learning_rate": 8e-05, "loss": 1.7369, "step": 5563 }, { "epoch": 0.6013184913001188, "grad_norm": 0.41611310839653015, "learning_rate": 8e-05, "loss": 1.6864, "step": 5564 }, { "epoch": 0.6014265643575056, "grad_norm": 0.40890011191368103, "learning_rate": 8e-05, "loss": 1.6397, "step": 5565 }, { "epoch": 0.6015346374148924, "grad_norm": 0.4269197881221771, "learning_rate": 8e-05, "loss": 1.7319, "step": 5566 }, { "epoch": 0.6016427104722792, "grad_norm": 0.3966192901134491, "learning_rate": 8e-05, "loss": 1.7092, "step": 5567 }, { "epoch": 0.601750783529666, "grad_norm": 0.4157271087169647, "learning_rate": 8e-05, "loss": 1.5161, "step": 5568 }, { "epoch": 0.6018588565870528, "grad_norm": 0.39074137806892395, "learning_rate": 8e-05, "loss": 1.5576, "step": 5569 }, { "epoch": 0.6019669296444397, "grad_norm": 0.4010371267795563, "learning_rate": 8e-05, "loss": 1.6178, "step": 5570 }, { "epoch": 0.6020750027018265, "grad_norm": 0.38313671946525574, "learning_rate": 8e-05, "loss": 1.7344, "step": 5571 }, { "epoch": 0.6021830757592133, "grad_norm": 0.41329509019851685, "learning_rate": 8e-05, "loss": 1.6359, "step": 5572 }, { "epoch": 0.6022911488166001, "grad_norm": 0.4341777265071869, "learning_rate": 8e-05, "loss": 1.7074, "step": 5573 }, { "epoch": 0.6023992218739869, "grad_norm": 0.42009955644607544, "learning_rate": 8e-05, "loss": 1.829, "step": 5574 }, { "epoch": 0.6025072949313736, "grad_norm": 0.39267870783805847, "learning_rate": 8e-05, "loss": 1.6626, "step": 5575 }, { "epoch": 0.6026153679887604, "grad_norm": 0.3903910219669342, "learning_rate": 8e-05, "loss": 1.6788, "step": 5576 }, { "epoch": 0.6027234410461472, "grad_norm": 0.4029584228992462, "learning_rate": 8e-05, "loss": 1.7467, "step": 5577 }, { "epoch": 0.602831514103534, "grad_norm": 0.40249747037887573, "learning_rate": 8e-05, "loss": 1.6888, "step": 5578 }, { "epoch": 0.6029395871609208, "grad_norm": 0.4511107802391052, "learning_rate": 8e-05, "loss": 1.7641, "step": 5579 }, { "epoch": 0.6030476602183076, "grad_norm": 0.3829243779182434, "learning_rate": 8e-05, "loss": 1.5017, "step": 5580 }, { "epoch": 0.6031557332756944, "grad_norm": 0.3954021632671356, "learning_rate": 8e-05, "loss": 1.6572, "step": 5581 }, { "epoch": 0.6032638063330812, "grad_norm": 0.4195549488067627, "learning_rate": 8e-05, "loss": 1.6922, "step": 5582 }, { "epoch": 0.603371879390468, "grad_norm": 0.44579875469207764, "learning_rate": 8e-05, "loss": 1.798, "step": 5583 }, { "epoch": 0.6034799524478548, "grad_norm": 0.42961937189102173, "learning_rate": 8e-05, "loss": 1.6561, "step": 5584 }, { "epoch": 0.6035880255052415, "grad_norm": 0.44027945399284363, "learning_rate": 8e-05, "loss": 1.5807, "step": 5585 }, { "epoch": 0.6036960985626283, "grad_norm": 0.41027164459228516, "learning_rate": 8e-05, "loss": 1.7927, "step": 5586 }, { "epoch": 0.6038041716200151, "grad_norm": 0.44223877787590027, "learning_rate": 8e-05, "loss": 1.8145, "step": 5587 }, { "epoch": 0.6039122446774019, "grad_norm": 0.4479760229587555, "learning_rate": 8e-05, "loss": 1.7708, "step": 5588 }, { "epoch": 0.6040203177347887, "grad_norm": 0.44671475887298584, "learning_rate": 8e-05, "loss": 1.7866, "step": 5589 }, { "epoch": 0.6041283907921755, "grad_norm": 0.4127994477748871, "learning_rate": 8e-05, "loss": 1.6093, "step": 5590 }, { "epoch": 0.6042364638495623, "grad_norm": 0.4356762170791626, "learning_rate": 8e-05, "loss": 1.685, "step": 5591 }, { "epoch": 0.6043445369069491, "grad_norm": 0.4292259216308594, "learning_rate": 8e-05, "loss": 1.5569, "step": 5592 }, { "epoch": 0.6044526099643359, "grad_norm": 0.4854232370853424, "learning_rate": 8e-05, "loss": 1.8519, "step": 5593 }, { "epoch": 0.6045606830217227, "grad_norm": 0.4225342273712158, "learning_rate": 8e-05, "loss": 1.8554, "step": 5594 }, { "epoch": 0.6046687560791095, "grad_norm": 0.5568528175354004, "learning_rate": 8e-05, "loss": 2.0601, "step": 5595 }, { "epoch": 0.6047768291364962, "grad_norm": 0.45186707377433777, "learning_rate": 8e-05, "loss": 1.7525, "step": 5596 }, { "epoch": 0.604884902193883, "grad_norm": 0.4047956168651581, "learning_rate": 8e-05, "loss": 1.6558, "step": 5597 }, { "epoch": 0.6049929752512698, "grad_norm": 0.3867219388484955, "learning_rate": 8e-05, "loss": 1.4988, "step": 5598 }, { "epoch": 0.6051010483086566, "grad_norm": 0.43252959847450256, "learning_rate": 8e-05, "loss": 1.4214, "step": 5599 }, { "epoch": 0.6052091213660434, "grad_norm": 0.4469071328639984, "learning_rate": 8e-05, "loss": 1.8918, "step": 5600 }, { "epoch": 0.6053171944234302, "grad_norm": 0.4018023908138275, "learning_rate": 8e-05, "loss": 1.6756, "step": 5601 }, { "epoch": 0.605425267480817, "grad_norm": 0.4045834243297577, "learning_rate": 8e-05, "loss": 1.5593, "step": 5602 }, { "epoch": 0.6055333405382038, "grad_norm": 0.4428858160972595, "learning_rate": 8e-05, "loss": 1.8375, "step": 5603 }, { "epoch": 0.6056414135955906, "grad_norm": 0.39916396141052246, "learning_rate": 8e-05, "loss": 1.6452, "step": 5604 }, { "epoch": 0.6057494866529775, "grad_norm": 0.4114767014980316, "learning_rate": 8e-05, "loss": 1.6721, "step": 5605 }, { "epoch": 0.6058575597103643, "grad_norm": 0.41199642419815063, "learning_rate": 8e-05, "loss": 1.7678, "step": 5606 }, { "epoch": 0.605965632767751, "grad_norm": 0.41069090366363525, "learning_rate": 8e-05, "loss": 1.5212, "step": 5607 }, { "epoch": 0.6060737058251378, "grad_norm": 0.42670223116874695, "learning_rate": 8e-05, "loss": 1.8181, "step": 5608 }, { "epoch": 0.6061817788825246, "grad_norm": 0.456539124250412, "learning_rate": 8e-05, "loss": 1.7159, "step": 5609 }, { "epoch": 0.6062898519399114, "grad_norm": 0.4061417877674103, "learning_rate": 8e-05, "loss": 1.7255, "step": 5610 }, { "epoch": 0.6063979249972982, "grad_norm": 0.45877355337142944, "learning_rate": 8e-05, "loss": 1.7268, "step": 5611 }, { "epoch": 0.606505998054685, "grad_norm": 0.41492971777915955, "learning_rate": 8e-05, "loss": 1.7781, "step": 5612 }, { "epoch": 0.6066140711120718, "grad_norm": 0.40113598108291626, "learning_rate": 8e-05, "loss": 1.6792, "step": 5613 }, { "epoch": 0.6067221441694586, "grad_norm": 0.4265323877334595, "learning_rate": 8e-05, "loss": 1.7544, "step": 5614 }, { "epoch": 0.6068302172268454, "grad_norm": 0.43082138895988464, "learning_rate": 8e-05, "loss": 1.5604, "step": 5615 }, { "epoch": 0.6069382902842322, "grad_norm": 0.4032224416732788, "learning_rate": 8e-05, "loss": 1.776, "step": 5616 }, { "epoch": 0.607046363341619, "grad_norm": 0.4472074806690216, "learning_rate": 8e-05, "loss": 1.6183, "step": 5617 }, { "epoch": 0.6071544363990057, "grad_norm": 0.41007599234580994, "learning_rate": 8e-05, "loss": 1.6999, "step": 5618 }, { "epoch": 0.6072625094563925, "grad_norm": 0.46281519532203674, "learning_rate": 8e-05, "loss": 1.6494, "step": 5619 }, { "epoch": 0.6073705825137793, "grad_norm": 0.4411652684211731, "learning_rate": 8e-05, "loss": 1.9231, "step": 5620 }, { "epoch": 0.6074786555711661, "grad_norm": 0.3987697958946228, "learning_rate": 8e-05, "loss": 1.5842, "step": 5621 }, { "epoch": 0.6075867286285529, "grad_norm": 0.41066330671310425, "learning_rate": 8e-05, "loss": 1.7389, "step": 5622 }, { "epoch": 0.6076948016859397, "grad_norm": 0.42821449041366577, "learning_rate": 8e-05, "loss": 1.6784, "step": 5623 }, { "epoch": 0.6078028747433265, "grad_norm": 0.4216649830341339, "learning_rate": 8e-05, "loss": 1.6187, "step": 5624 }, { "epoch": 0.6079109478007133, "grad_norm": 0.42534059286117554, "learning_rate": 8e-05, "loss": 1.6479, "step": 5625 }, { "epoch": 0.6080190208581001, "grad_norm": 0.44120925664901733, "learning_rate": 8e-05, "loss": 1.8128, "step": 5626 }, { "epoch": 0.6081270939154869, "grad_norm": 0.4123339354991913, "learning_rate": 8e-05, "loss": 1.64, "step": 5627 }, { "epoch": 0.6082351669728736, "grad_norm": 0.500092089176178, "learning_rate": 8e-05, "loss": 1.7878, "step": 5628 }, { "epoch": 0.6083432400302604, "grad_norm": 0.41842466592788696, "learning_rate": 8e-05, "loss": 1.7727, "step": 5629 }, { "epoch": 0.6084513130876472, "grad_norm": 0.4367406964302063, "learning_rate": 8e-05, "loss": 1.6617, "step": 5630 }, { "epoch": 0.608559386145034, "grad_norm": 0.48225516080856323, "learning_rate": 8e-05, "loss": 1.906, "step": 5631 }, { "epoch": 0.6086674592024208, "grad_norm": 0.40696126222610474, "learning_rate": 8e-05, "loss": 1.5374, "step": 5632 }, { "epoch": 0.6087755322598076, "grad_norm": 0.43131428956985474, "learning_rate": 8e-05, "loss": 1.7023, "step": 5633 }, { "epoch": 0.6088836053171944, "grad_norm": 0.41503649950027466, "learning_rate": 8e-05, "loss": 1.6495, "step": 5634 }, { "epoch": 0.6089916783745812, "grad_norm": 0.4401205778121948, "learning_rate": 8e-05, "loss": 1.7384, "step": 5635 }, { "epoch": 0.609099751431968, "grad_norm": 0.4417973756790161, "learning_rate": 8e-05, "loss": 1.6984, "step": 5636 }, { "epoch": 0.6092078244893548, "grad_norm": 0.42758750915527344, "learning_rate": 8e-05, "loss": 1.8865, "step": 5637 }, { "epoch": 0.6093158975467416, "grad_norm": 0.3931749165058136, "learning_rate": 8e-05, "loss": 1.6197, "step": 5638 }, { "epoch": 0.6094239706041283, "grad_norm": 0.4368842542171478, "learning_rate": 8e-05, "loss": 1.7949, "step": 5639 }, { "epoch": 0.6095320436615151, "grad_norm": 0.47171565890312195, "learning_rate": 8e-05, "loss": 1.9155, "step": 5640 }, { "epoch": 0.609640116718902, "grad_norm": 0.4479542374610901, "learning_rate": 8e-05, "loss": 1.8276, "step": 5641 }, { "epoch": 0.6097481897762888, "grad_norm": 0.40911000967025757, "learning_rate": 8e-05, "loss": 1.696, "step": 5642 }, { "epoch": 0.6098562628336756, "grad_norm": 0.43679261207580566, "learning_rate": 8e-05, "loss": 1.6555, "step": 5643 }, { "epoch": 0.6099643358910624, "grad_norm": 0.42374613881111145, "learning_rate": 8e-05, "loss": 1.7698, "step": 5644 }, { "epoch": 0.6100724089484492, "grad_norm": 0.41835370659828186, "learning_rate": 8e-05, "loss": 1.6622, "step": 5645 }, { "epoch": 0.610180482005836, "grad_norm": 0.42975136637687683, "learning_rate": 8e-05, "loss": 1.7731, "step": 5646 }, { "epoch": 0.6102885550632228, "grad_norm": 0.4307366609573364, "learning_rate": 8e-05, "loss": 1.7752, "step": 5647 }, { "epoch": 0.6103966281206096, "grad_norm": 0.4630015790462494, "learning_rate": 8e-05, "loss": 1.7038, "step": 5648 }, { "epoch": 0.6105047011779964, "grad_norm": 0.45784279704093933, "learning_rate": 8e-05, "loss": 1.8174, "step": 5649 }, { "epoch": 0.6106127742353831, "grad_norm": 0.43445196747779846, "learning_rate": 8e-05, "loss": 1.8712, "step": 5650 }, { "epoch": 0.6107208472927699, "grad_norm": 0.39535608887672424, "learning_rate": 8e-05, "loss": 1.4865, "step": 5651 }, { "epoch": 0.6108289203501567, "grad_norm": 0.3966144323348999, "learning_rate": 8e-05, "loss": 1.7372, "step": 5652 }, { "epoch": 0.6109369934075435, "grad_norm": 0.4390961825847626, "learning_rate": 8e-05, "loss": 1.8847, "step": 5653 }, { "epoch": 0.6110450664649303, "grad_norm": 0.3831745982170105, "learning_rate": 8e-05, "loss": 1.438, "step": 5654 }, { "epoch": 0.6111531395223171, "grad_norm": 0.42287588119506836, "learning_rate": 8e-05, "loss": 1.7079, "step": 5655 }, { "epoch": 0.6112612125797039, "grad_norm": 0.4569057822227478, "learning_rate": 8e-05, "loss": 1.8525, "step": 5656 }, { "epoch": 0.6113692856370907, "grad_norm": 0.3908258378505707, "learning_rate": 8e-05, "loss": 1.6774, "step": 5657 }, { "epoch": 0.6114773586944775, "grad_norm": 0.4081214964389801, "learning_rate": 8e-05, "loss": 1.6763, "step": 5658 }, { "epoch": 0.6115854317518643, "grad_norm": 0.39823055267333984, "learning_rate": 8e-05, "loss": 1.5901, "step": 5659 }, { "epoch": 0.611693504809251, "grad_norm": 0.41250815987586975, "learning_rate": 8e-05, "loss": 1.5692, "step": 5660 }, { "epoch": 0.6118015778666378, "grad_norm": 0.45517972111701965, "learning_rate": 8e-05, "loss": 1.8858, "step": 5661 }, { "epoch": 0.6119096509240246, "grad_norm": 0.4603644609451294, "learning_rate": 8e-05, "loss": 1.845, "step": 5662 }, { "epoch": 0.6120177239814114, "grad_norm": 0.426254004240036, "learning_rate": 8e-05, "loss": 1.7627, "step": 5663 }, { "epoch": 0.6121257970387982, "grad_norm": 0.403541624546051, "learning_rate": 8e-05, "loss": 1.7018, "step": 5664 }, { "epoch": 0.612233870096185, "grad_norm": 0.3930877149105072, "learning_rate": 8e-05, "loss": 1.5413, "step": 5665 }, { "epoch": 0.6123419431535718, "grad_norm": 0.4229961037635803, "learning_rate": 8e-05, "loss": 1.8343, "step": 5666 }, { "epoch": 0.6124500162109586, "grad_norm": 0.4302409291267395, "learning_rate": 8e-05, "loss": 1.7411, "step": 5667 }, { "epoch": 0.6125580892683454, "grad_norm": 0.4032720625400543, "learning_rate": 8e-05, "loss": 1.5035, "step": 5668 }, { "epoch": 0.6126661623257322, "grad_norm": 0.4326929748058319, "learning_rate": 8e-05, "loss": 1.685, "step": 5669 }, { "epoch": 0.612774235383119, "grad_norm": 0.416983425617218, "learning_rate": 8e-05, "loss": 1.6782, "step": 5670 }, { "epoch": 0.6128823084405057, "grad_norm": 0.4509740471839905, "learning_rate": 8e-05, "loss": 1.6032, "step": 5671 }, { "epoch": 0.6129903814978925, "grad_norm": 0.39977389574050903, "learning_rate": 8e-05, "loss": 1.6711, "step": 5672 }, { "epoch": 0.6130984545552793, "grad_norm": 0.43511274456977844, "learning_rate": 8e-05, "loss": 1.6911, "step": 5673 }, { "epoch": 0.6132065276126661, "grad_norm": 0.4287088215351105, "learning_rate": 8e-05, "loss": 1.807, "step": 5674 }, { "epoch": 0.613314600670053, "grad_norm": 0.4657456874847412, "learning_rate": 8e-05, "loss": 1.7497, "step": 5675 }, { "epoch": 0.6134226737274397, "grad_norm": 0.42656806111335754, "learning_rate": 8e-05, "loss": 1.813, "step": 5676 }, { "epoch": 0.6135307467848266, "grad_norm": 0.3862333297729492, "learning_rate": 8e-05, "loss": 1.6035, "step": 5677 }, { "epoch": 0.6136388198422134, "grad_norm": 0.40622854232788086, "learning_rate": 8e-05, "loss": 1.6475, "step": 5678 }, { "epoch": 0.6137468928996002, "grad_norm": 0.42562615871429443, "learning_rate": 8e-05, "loss": 1.6222, "step": 5679 }, { "epoch": 0.613854965956987, "grad_norm": 0.39238396286964417, "learning_rate": 8e-05, "loss": 1.5806, "step": 5680 }, { "epoch": 0.6139630390143738, "grad_norm": 0.4338913559913635, "learning_rate": 8e-05, "loss": 1.6155, "step": 5681 }, { "epoch": 0.6140711120717605, "grad_norm": 0.44086387753486633, "learning_rate": 8e-05, "loss": 1.7364, "step": 5682 }, { "epoch": 0.6141791851291473, "grad_norm": 0.43404102325439453, "learning_rate": 8e-05, "loss": 1.8058, "step": 5683 }, { "epoch": 0.6142872581865341, "grad_norm": 0.4091910123825073, "learning_rate": 8e-05, "loss": 1.6101, "step": 5684 }, { "epoch": 0.6143953312439209, "grad_norm": 0.41352197527885437, "learning_rate": 8e-05, "loss": 1.8846, "step": 5685 }, { "epoch": 0.6145034043013077, "grad_norm": 0.42651283740997314, "learning_rate": 8e-05, "loss": 1.7239, "step": 5686 }, { "epoch": 0.6146114773586945, "grad_norm": 0.4622993767261505, "learning_rate": 8e-05, "loss": 1.8527, "step": 5687 }, { "epoch": 0.6147195504160813, "grad_norm": 0.4348565340042114, "learning_rate": 8e-05, "loss": 1.7455, "step": 5688 }, { "epoch": 0.6148276234734681, "grad_norm": 0.4010656177997589, "learning_rate": 8e-05, "loss": 1.6054, "step": 5689 }, { "epoch": 0.6149356965308549, "grad_norm": 0.4157104194164276, "learning_rate": 8e-05, "loss": 1.705, "step": 5690 }, { "epoch": 0.6150437695882417, "grad_norm": 0.3911777138710022, "learning_rate": 8e-05, "loss": 1.5185, "step": 5691 }, { "epoch": 0.6151518426456285, "grad_norm": 0.39344099164009094, "learning_rate": 8e-05, "loss": 1.7171, "step": 5692 }, { "epoch": 0.6152599157030152, "grad_norm": 0.4162250757217407, "learning_rate": 8e-05, "loss": 1.6939, "step": 5693 }, { "epoch": 0.615367988760402, "grad_norm": 0.40147605538368225, "learning_rate": 8e-05, "loss": 1.6942, "step": 5694 }, { "epoch": 0.6154760618177888, "grad_norm": 0.40599560737609863, "learning_rate": 8e-05, "loss": 1.5726, "step": 5695 }, { "epoch": 0.6155841348751756, "grad_norm": 0.41977331042289734, "learning_rate": 8e-05, "loss": 1.695, "step": 5696 }, { "epoch": 0.6156922079325624, "grad_norm": 0.39120498299598694, "learning_rate": 8e-05, "loss": 1.6853, "step": 5697 }, { "epoch": 0.6158002809899492, "grad_norm": 0.4144049286842346, "learning_rate": 8e-05, "loss": 1.7999, "step": 5698 }, { "epoch": 0.615908354047336, "grad_norm": 0.3860338628292084, "learning_rate": 8e-05, "loss": 1.5688, "step": 5699 }, { "epoch": 0.6160164271047228, "grad_norm": 0.4165055751800537, "learning_rate": 8e-05, "loss": 1.5805, "step": 5700 }, { "epoch": 0.6161245001621096, "grad_norm": 0.4285001754760742, "learning_rate": 8e-05, "loss": 1.7881, "step": 5701 }, { "epoch": 0.6162325732194964, "grad_norm": 0.3942855894565582, "learning_rate": 8e-05, "loss": 1.5927, "step": 5702 }, { "epoch": 0.6163406462768831, "grad_norm": 0.42704445123672485, "learning_rate": 8e-05, "loss": 1.704, "step": 5703 }, { "epoch": 0.6164487193342699, "grad_norm": 0.4244515299797058, "learning_rate": 8e-05, "loss": 1.7124, "step": 5704 }, { "epoch": 0.6165567923916567, "grad_norm": 0.40719887614250183, "learning_rate": 8e-05, "loss": 1.4892, "step": 5705 }, { "epoch": 0.6166648654490435, "grad_norm": 0.40813493728637695, "learning_rate": 8e-05, "loss": 1.7363, "step": 5706 }, { "epoch": 0.6167729385064303, "grad_norm": 0.40764933824539185, "learning_rate": 8e-05, "loss": 1.5631, "step": 5707 }, { "epoch": 0.6168810115638171, "grad_norm": 0.4487496614456177, "learning_rate": 8e-05, "loss": 1.8091, "step": 5708 }, { "epoch": 0.6169890846212039, "grad_norm": 0.39770346879959106, "learning_rate": 8e-05, "loss": 1.6506, "step": 5709 }, { "epoch": 0.6170971576785907, "grad_norm": 0.4144371747970581, "learning_rate": 8e-05, "loss": 1.6373, "step": 5710 }, { "epoch": 0.6172052307359776, "grad_norm": 0.44960346817970276, "learning_rate": 8e-05, "loss": 1.7786, "step": 5711 }, { "epoch": 0.6173133037933644, "grad_norm": 0.43939679861068726, "learning_rate": 8e-05, "loss": 1.7627, "step": 5712 }, { "epoch": 0.6174213768507512, "grad_norm": 0.42717891931533813, "learning_rate": 8e-05, "loss": 1.7354, "step": 5713 }, { "epoch": 0.6175294499081379, "grad_norm": 0.4149041175842285, "learning_rate": 8e-05, "loss": 1.6113, "step": 5714 }, { "epoch": 0.6176375229655247, "grad_norm": 0.4046395421028137, "learning_rate": 8e-05, "loss": 1.5454, "step": 5715 }, { "epoch": 0.6177455960229115, "grad_norm": 0.4560054540634155, "learning_rate": 8e-05, "loss": 1.7716, "step": 5716 }, { "epoch": 0.6178536690802983, "grad_norm": 0.4470860958099365, "learning_rate": 8e-05, "loss": 1.7308, "step": 5717 }, { "epoch": 0.6179617421376851, "grad_norm": 0.4559294879436493, "learning_rate": 8e-05, "loss": 1.8195, "step": 5718 }, { "epoch": 0.6180698151950719, "grad_norm": 0.46014729142189026, "learning_rate": 8e-05, "loss": 1.6874, "step": 5719 }, { "epoch": 0.6181778882524587, "grad_norm": 0.43334531784057617, "learning_rate": 8e-05, "loss": 1.6934, "step": 5720 }, { "epoch": 0.6182859613098455, "grad_norm": 0.3951547145843506, "learning_rate": 8e-05, "loss": 1.6405, "step": 5721 }, { "epoch": 0.6183940343672323, "grad_norm": 0.4764918088912964, "learning_rate": 8e-05, "loss": 1.9216, "step": 5722 }, { "epoch": 0.6185021074246191, "grad_norm": 0.43205007910728455, "learning_rate": 8e-05, "loss": 1.7716, "step": 5723 }, { "epoch": 0.6186101804820059, "grad_norm": 0.4268185496330261, "learning_rate": 8e-05, "loss": 1.9038, "step": 5724 }, { "epoch": 0.6187182535393926, "grad_norm": 0.4258314371109009, "learning_rate": 8e-05, "loss": 1.6863, "step": 5725 }, { "epoch": 0.6188263265967794, "grad_norm": 0.40796738862991333, "learning_rate": 8e-05, "loss": 1.7046, "step": 5726 }, { "epoch": 0.6189343996541662, "grad_norm": 0.38076791167259216, "learning_rate": 8e-05, "loss": 1.5298, "step": 5727 }, { "epoch": 0.619042472711553, "grad_norm": 0.47369658946990967, "learning_rate": 8e-05, "loss": 1.7461, "step": 5728 }, { "epoch": 0.6191505457689398, "grad_norm": 0.44274717569351196, "learning_rate": 8e-05, "loss": 1.8943, "step": 5729 }, { "epoch": 0.6192586188263266, "grad_norm": 0.4361170530319214, "learning_rate": 8e-05, "loss": 1.853, "step": 5730 }, { "epoch": 0.6193666918837134, "grad_norm": 0.4226546585559845, "learning_rate": 8e-05, "loss": 1.6227, "step": 5731 }, { "epoch": 0.6194747649411002, "grad_norm": 0.3840100169181824, "learning_rate": 8e-05, "loss": 1.5828, "step": 5732 }, { "epoch": 0.619582837998487, "grad_norm": 0.410635381937027, "learning_rate": 8e-05, "loss": 1.5974, "step": 5733 }, { "epoch": 0.6196909110558738, "grad_norm": 0.418998122215271, "learning_rate": 8e-05, "loss": 1.8365, "step": 5734 }, { "epoch": 0.6197989841132605, "grad_norm": 0.44556674361228943, "learning_rate": 8e-05, "loss": 1.8595, "step": 5735 }, { "epoch": 0.6199070571706473, "grad_norm": 0.4061774015426636, "learning_rate": 8e-05, "loss": 1.6039, "step": 5736 }, { "epoch": 0.6200151302280341, "grad_norm": 0.42079100012779236, "learning_rate": 8e-05, "loss": 1.8145, "step": 5737 }, { "epoch": 0.6201232032854209, "grad_norm": 0.426146000623703, "learning_rate": 8e-05, "loss": 1.6699, "step": 5738 }, { "epoch": 0.6202312763428077, "grad_norm": 0.41415050625801086, "learning_rate": 8e-05, "loss": 1.8142, "step": 5739 }, { "epoch": 0.6203393494001945, "grad_norm": 0.4181089401245117, "learning_rate": 8e-05, "loss": 1.5438, "step": 5740 }, { "epoch": 0.6204474224575813, "grad_norm": 0.3856123387813568, "learning_rate": 8e-05, "loss": 1.6409, "step": 5741 }, { "epoch": 0.6205554955149681, "grad_norm": 0.3778034448623657, "learning_rate": 8e-05, "loss": 1.369, "step": 5742 }, { "epoch": 0.6206635685723549, "grad_norm": 0.4220992624759674, "learning_rate": 8e-05, "loss": 1.7196, "step": 5743 }, { "epoch": 0.6207716416297417, "grad_norm": 0.4229843318462372, "learning_rate": 8e-05, "loss": 1.6235, "step": 5744 }, { "epoch": 0.6208797146871285, "grad_norm": 0.3773055374622345, "learning_rate": 8e-05, "loss": 1.6334, "step": 5745 }, { "epoch": 0.6209877877445152, "grad_norm": 0.4290542006492615, "learning_rate": 8e-05, "loss": 1.6119, "step": 5746 }, { "epoch": 0.621095860801902, "grad_norm": 0.433635413646698, "learning_rate": 8e-05, "loss": 1.7372, "step": 5747 }, { "epoch": 0.6212039338592888, "grad_norm": 0.4553503096103668, "learning_rate": 8e-05, "loss": 1.8622, "step": 5748 }, { "epoch": 0.6213120069166757, "grad_norm": 0.39354079961776733, "learning_rate": 8e-05, "loss": 1.662, "step": 5749 }, { "epoch": 0.6214200799740625, "grad_norm": 0.409074068069458, "learning_rate": 8e-05, "loss": 1.7102, "step": 5750 }, { "epoch": 0.6215281530314493, "grad_norm": 0.38076263666152954, "learning_rate": 8e-05, "loss": 1.6518, "step": 5751 }, { "epoch": 0.6216362260888361, "grad_norm": 0.44247356057167053, "learning_rate": 8e-05, "loss": 1.7431, "step": 5752 }, { "epoch": 0.6217442991462229, "grad_norm": 0.46258601546287537, "learning_rate": 8e-05, "loss": 1.9798, "step": 5753 }, { "epoch": 0.6218523722036097, "grad_norm": 0.4306080639362335, "learning_rate": 8e-05, "loss": 1.7223, "step": 5754 }, { "epoch": 0.6219604452609965, "grad_norm": 0.4006788432598114, "learning_rate": 8e-05, "loss": 1.6893, "step": 5755 }, { "epoch": 0.6220685183183833, "grad_norm": 0.4497901201248169, "learning_rate": 8e-05, "loss": 1.6622, "step": 5756 }, { "epoch": 0.62217659137577, "grad_norm": 0.41648179292678833, "learning_rate": 8e-05, "loss": 1.7597, "step": 5757 }, { "epoch": 0.6222846644331568, "grad_norm": 0.390455424785614, "learning_rate": 8e-05, "loss": 1.7141, "step": 5758 }, { "epoch": 0.6223927374905436, "grad_norm": 0.40265893936157227, "learning_rate": 8e-05, "loss": 1.5433, "step": 5759 }, { "epoch": 0.6225008105479304, "grad_norm": 0.45590636134147644, "learning_rate": 8e-05, "loss": 1.8326, "step": 5760 }, { "epoch": 0.6226088836053172, "grad_norm": 0.43206387758255005, "learning_rate": 8e-05, "loss": 1.6969, "step": 5761 }, { "epoch": 0.622716956662704, "grad_norm": 0.42712536454200745, "learning_rate": 8e-05, "loss": 1.7298, "step": 5762 }, { "epoch": 0.6228250297200908, "grad_norm": 0.41202470660209656, "learning_rate": 8e-05, "loss": 1.7807, "step": 5763 }, { "epoch": 0.6229331027774776, "grad_norm": 0.4218021035194397, "learning_rate": 8e-05, "loss": 1.5666, "step": 5764 }, { "epoch": 0.6230411758348644, "grad_norm": 0.39102327823638916, "learning_rate": 8e-05, "loss": 1.6829, "step": 5765 }, { "epoch": 0.6231492488922512, "grad_norm": 0.3961150050163269, "learning_rate": 8e-05, "loss": 1.6329, "step": 5766 }, { "epoch": 0.623257321949638, "grad_norm": 0.42224088311195374, "learning_rate": 8e-05, "loss": 1.7846, "step": 5767 }, { "epoch": 0.6233653950070247, "grad_norm": 0.4417797327041626, "learning_rate": 8e-05, "loss": 1.5063, "step": 5768 }, { "epoch": 0.6234734680644115, "grad_norm": 0.40327778458595276, "learning_rate": 8e-05, "loss": 1.6164, "step": 5769 }, { "epoch": 0.6235815411217983, "grad_norm": 0.4157952070236206, "learning_rate": 8e-05, "loss": 1.5256, "step": 5770 }, { "epoch": 0.6236896141791851, "grad_norm": 0.4010724425315857, "learning_rate": 8e-05, "loss": 1.6799, "step": 5771 }, { "epoch": 0.6237976872365719, "grad_norm": 0.46424973011016846, "learning_rate": 8e-05, "loss": 1.7961, "step": 5772 }, { "epoch": 0.6239057602939587, "grad_norm": 0.4089820683002472, "learning_rate": 8e-05, "loss": 1.577, "step": 5773 }, { "epoch": 0.6240138333513455, "grad_norm": 0.4485880136489868, "learning_rate": 8e-05, "loss": 1.7948, "step": 5774 }, { "epoch": 0.6241219064087323, "grad_norm": 0.46849390864372253, "learning_rate": 8e-05, "loss": 1.77, "step": 5775 }, { "epoch": 0.6242299794661191, "grad_norm": 0.40395665168762207, "learning_rate": 8e-05, "loss": 1.8019, "step": 5776 }, { "epoch": 0.6243380525235059, "grad_norm": 0.43334877490997314, "learning_rate": 8e-05, "loss": 1.6247, "step": 5777 }, { "epoch": 0.6244461255808926, "grad_norm": 0.4426073431968689, "learning_rate": 8e-05, "loss": 1.6143, "step": 5778 }, { "epoch": 0.6245541986382794, "grad_norm": 0.42289671301841736, "learning_rate": 8e-05, "loss": 1.7873, "step": 5779 }, { "epoch": 0.6246622716956662, "grad_norm": 0.3978530168533325, "learning_rate": 8e-05, "loss": 1.7271, "step": 5780 }, { "epoch": 0.624770344753053, "grad_norm": 0.39185091853141785, "learning_rate": 8e-05, "loss": 1.6214, "step": 5781 }, { "epoch": 0.6248784178104398, "grad_norm": 0.37814491987228394, "learning_rate": 8e-05, "loss": 1.5137, "step": 5782 }, { "epoch": 0.6249864908678266, "grad_norm": 0.4660125970840454, "learning_rate": 8e-05, "loss": 1.7786, "step": 5783 }, { "epoch": 0.6250945639252135, "grad_norm": 0.41629624366760254, "learning_rate": 8e-05, "loss": 1.7168, "step": 5784 }, { "epoch": 0.6252026369826003, "grad_norm": 0.40925827622413635, "learning_rate": 8e-05, "loss": 1.4092, "step": 5785 }, { "epoch": 0.6253107100399871, "grad_norm": 0.41336894035339355, "learning_rate": 8e-05, "loss": 1.7053, "step": 5786 }, { "epoch": 0.6254187830973739, "grad_norm": 0.43133464455604553, "learning_rate": 8e-05, "loss": 1.7367, "step": 5787 }, { "epoch": 0.6255268561547607, "grad_norm": 0.5405160784721375, "learning_rate": 8e-05, "loss": 1.5703, "step": 5788 }, { "epoch": 0.6256349292121474, "grad_norm": 0.4014160633087158, "learning_rate": 8e-05, "loss": 1.7157, "step": 5789 }, { "epoch": 0.6257430022695342, "grad_norm": 0.3916128873825073, "learning_rate": 8e-05, "loss": 1.7409, "step": 5790 }, { "epoch": 0.625851075326921, "grad_norm": 0.44866815209388733, "learning_rate": 8e-05, "loss": 1.7691, "step": 5791 }, { "epoch": 0.6259591483843078, "grad_norm": 0.3947921395301819, "learning_rate": 8e-05, "loss": 1.6201, "step": 5792 }, { "epoch": 0.6260672214416946, "grad_norm": 0.4754680395126343, "learning_rate": 8e-05, "loss": 1.6178, "step": 5793 }, { "epoch": 0.6261752944990814, "grad_norm": 0.4054376184940338, "learning_rate": 8e-05, "loss": 1.6968, "step": 5794 }, { "epoch": 0.6262833675564682, "grad_norm": 0.4367932379245758, "learning_rate": 8e-05, "loss": 1.5732, "step": 5795 }, { "epoch": 0.626391440613855, "grad_norm": 0.4554263651371002, "learning_rate": 8e-05, "loss": 1.6554, "step": 5796 }, { "epoch": 0.6264995136712418, "grad_norm": 0.4203793406486511, "learning_rate": 8e-05, "loss": 1.6125, "step": 5797 }, { "epoch": 0.6266075867286286, "grad_norm": 0.4763276278972626, "learning_rate": 8e-05, "loss": 1.9069, "step": 5798 }, { "epoch": 0.6267156597860154, "grad_norm": 0.40524324774742126, "learning_rate": 8e-05, "loss": 1.6596, "step": 5799 }, { "epoch": 0.6268237328434021, "grad_norm": 0.40505340695381165, "learning_rate": 8e-05, "loss": 1.685, "step": 5800 }, { "epoch": 0.6269318059007889, "grad_norm": 0.4406121075153351, "learning_rate": 8e-05, "loss": 1.638, "step": 5801 }, { "epoch": 0.6270398789581757, "grad_norm": 0.46213680505752563, "learning_rate": 8e-05, "loss": 1.7627, "step": 5802 }, { "epoch": 0.6271479520155625, "grad_norm": 0.41784486174583435, "learning_rate": 8e-05, "loss": 1.5897, "step": 5803 }, { "epoch": 0.6272560250729493, "grad_norm": 0.42149707674980164, "learning_rate": 8e-05, "loss": 1.6813, "step": 5804 }, { "epoch": 0.6273640981303361, "grad_norm": 0.47300130128860474, "learning_rate": 8e-05, "loss": 1.6431, "step": 5805 }, { "epoch": 0.6274721711877229, "grad_norm": 0.4764530062675476, "learning_rate": 8e-05, "loss": 1.8681, "step": 5806 }, { "epoch": 0.6275802442451097, "grad_norm": 0.3965219557285309, "learning_rate": 8e-05, "loss": 1.4941, "step": 5807 }, { "epoch": 0.6276883173024965, "grad_norm": 0.4540013372898102, "learning_rate": 8e-05, "loss": 1.7236, "step": 5808 }, { "epoch": 0.6277963903598833, "grad_norm": 0.394480437040329, "learning_rate": 8e-05, "loss": 1.7182, "step": 5809 }, { "epoch": 0.62790446341727, "grad_norm": 0.4563153386116028, "learning_rate": 8e-05, "loss": 1.6679, "step": 5810 }, { "epoch": 0.6280125364746568, "grad_norm": 0.44134581089019775, "learning_rate": 8e-05, "loss": 1.6187, "step": 5811 }, { "epoch": 0.6281206095320436, "grad_norm": 0.4445981979370117, "learning_rate": 8e-05, "loss": 1.8399, "step": 5812 }, { "epoch": 0.6282286825894304, "grad_norm": 0.4003884196281433, "learning_rate": 8e-05, "loss": 1.6945, "step": 5813 }, { "epoch": 0.6283367556468172, "grad_norm": 0.4417981505393982, "learning_rate": 8e-05, "loss": 1.8548, "step": 5814 }, { "epoch": 0.628444828704204, "grad_norm": 0.4024435877799988, "learning_rate": 8e-05, "loss": 1.6563, "step": 5815 }, { "epoch": 0.6285529017615908, "grad_norm": 0.4311235249042511, "learning_rate": 8e-05, "loss": 1.6971, "step": 5816 }, { "epoch": 0.6286609748189776, "grad_norm": 0.4685044586658478, "learning_rate": 8e-05, "loss": 1.7731, "step": 5817 }, { "epoch": 0.6287690478763645, "grad_norm": 0.44598692655563354, "learning_rate": 8e-05, "loss": 1.8071, "step": 5818 }, { "epoch": 0.6288771209337513, "grad_norm": 0.4459400475025177, "learning_rate": 8e-05, "loss": 1.6525, "step": 5819 }, { "epoch": 0.6289851939911381, "grad_norm": 0.4133986830711365, "learning_rate": 8e-05, "loss": 1.644, "step": 5820 }, { "epoch": 0.6290932670485248, "grad_norm": 0.3954031467437744, "learning_rate": 8e-05, "loss": 1.5787, "step": 5821 }, { "epoch": 0.6292013401059116, "grad_norm": 0.46168646216392517, "learning_rate": 8e-05, "loss": 1.792, "step": 5822 }, { "epoch": 0.6293094131632984, "grad_norm": 0.44147804379463196, "learning_rate": 8e-05, "loss": 1.4824, "step": 5823 }, { "epoch": 0.6294174862206852, "grad_norm": 0.39809828996658325, "learning_rate": 8e-05, "loss": 1.6129, "step": 5824 }, { "epoch": 0.629525559278072, "grad_norm": 0.3972349464893341, "learning_rate": 8e-05, "loss": 1.6469, "step": 5825 }, { "epoch": 0.6296336323354588, "grad_norm": 0.4885788559913635, "learning_rate": 8e-05, "loss": 1.7081, "step": 5826 }, { "epoch": 0.6297417053928456, "grad_norm": 0.43848690390586853, "learning_rate": 8e-05, "loss": 1.8034, "step": 5827 }, { "epoch": 0.6298497784502324, "grad_norm": 0.4647013247013092, "learning_rate": 8e-05, "loss": 1.7635, "step": 5828 }, { "epoch": 0.6299578515076192, "grad_norm": 0.40279915928840637, "learning_rate": 8e-05, "loss": 1.5837, "step": 5829 }, { "epoch": 0.630065924565006, "grad_norm": 0.4532300531864166, "learning_rate": 8e-05, "loss": 1.7288, "step": 5830 }, { "epoch": 0.6301739976223928, "grad_norm": 0.4230743646621704, "learning_rate": 8e-05, "loss": 1.775, "step": 5831 }, { "epoch": 0.6302820706797795, "grad_norm": 0.4238855242729187, "learning_rate": 8e-05, "loss": 1.6748, "step": 5832 }, { "epoch": 0.6303901437371663, "grad_norm": 0.42089831829071045, "learning_rate": 8e-05, "loss": 1.7349, "step": 5833 }, { "epoch": 0.6304982167945531, "grad_norm": 0.42921558022499084, "learning_rate": 8e-05, "loss": 1.8772, "step": 5834 }, { "epoch": 0.6306062898519399, "grad_norm": 0.426667720079422, "learning_rate": 8e-05, "loss": 1.7373, "step": 5835 }, { "epoch": 0.6307143629093267, "grad_norm": 0.43521344661712646, "learning_rate": 8e-05, "loss": 1.6367, "step": 5836 }, { "epoch": 0.6308224359667135, "grad_norm": 0.4293736219406128, "learning_rate": 8e-05, "loss": 1.4526, "step": 5837 }, { "epoch": 0.6309305090241003, "grad_norm": 0.40097296237945557, "learning_rate": 8e-05, "loss": 1.7579, "step": 5838 }, { "epoch": 0.6310385820814871, "grad_norm": 0.4617515802383423, "learning_rate": 8e-05, "loss": 1.7811, "step": 5839 }, { "epoch": 0.6311466551388739, "grad_norm": 0.4587014615535736, "learning_rate": 8e-05, "loss": 1.7128, "step": 5840 }, { "epoch": 0.6312547281962607, "grad_norm": 0.4044555425643921, "learning_rate": 8e-05, "loss": 1.7054, "step": 5841 }, { "epoch": 0.6313628012536475, "grad_norm": 0.4363962411880493, "learning_rate": 8e-05, "loss": 1.572, "step": 5842 }, { "epoch": 0.6314708743110342, "grad_norm": 0.4211708903312683, "learning_rate": 8e-05, "loss": 1.6977, "step": 5843 }, { "epoch": 0.631578947368421, "grad_norm": 0.42466625571250916, "learning_rate": 8e-05, "loss": 1.7098, "step": 5844 }, { "epoch": 0.6316870204258078, "grad_norm": 0.444354772567749, "learning_rate": 8e-05, "loss": 1.7662, "step": 5845 }, { "epoch": 0.6317950934831946, "grad_norm": 0.42519158124923706, "learning_rate": 8e-05, "loss": 1.3476, "step": 5846 }, { "epoch": 0.6319031665405814, "grad_norm": 0.45832714438438416, "learning_rate": 8e-05, "loss": 1.6904, "step": 5847 }, { "epoch": 0.6320112395979682, "grad_norm": 0.41959547996520996, "learning_rate": 8e-05, "loss": 1.75, "step": 5848 }, { "epoch": 0.632119312655355, "grad_norm": 0.4385189712047577, "learning_rate": 8e-05, "loss": 1.6764, "step": 5849 }, { "epoch": 0.6322273857127418, "grad_norm": 0.4105124771595001, "learning_rate": 8e-05, "loss": 1.5979, "step": 5850 }, { "epoch": 0.6323354587701286, "grad_norm": 0.4027622640132904, "learning_rate": 8e-05, "loss": 1.5868, "step": 5851 }, { "epoch": 0.6324435318275154, "grad_norm": 0.5007439851760864, "learning_rate": 8e-05, "loss": 1.8078, "step": 5852 }, { "epoch": 0.6325516048849021, "grad_norm": 0.41511887311935425, "learning_rate": 8e-05, "loss": 1.5848, "step": 5853 }, { "epoch": 0.632659677942289, "grad_norm": 0.43523499369621277, "learning_rate": 8e-05, "loss": 1.687, "step": 5854 }, { "epoch": 0.6327677509996757, "grad_norm": 0.4846849739551544, "learning_rate": 8e-05, "loss": 1.7754, "step": 5855 }, { "epoch": 0.6328758240570626, "grad_norm": 0.44204580783843994, "learning_rate": 8e-05, "loss": 1.7688, "step": 5856 }, { "epoch": 0.6329838971144494, "grad_norm": 0.3878631889820099, "learning_rate": 8e-05, "loss": 1.5956, "step": 5857 }, { "epoch": 0.6330919701718362, "grad_norm": 0.4276898205280304, "learning_rate": 8e-05, "loss": 1.7779, "step": 5858 }, { "epoch": 0.633200043229223, "grad_norm": 0.41155996918678284, "learning_rate": 8e-05, "loss": 1.7371, "step": 5859 }, { "epoch": 0.6333081162866098, "grad_norm": 0.41365593671798706, "learning_rate": 8e-05, "loss": 1.6848, "step": 5860 }, { "epoch": 0.6334161893439966, "grad_norm": 0.4193393588066101, "learning_rate": 8e-05, "loss": 1.7067, "step": 5861 }, { "epoch": 0.6335242624013834, "grad_norm": 0.4425993859767914, "learning_rate": 8e-05, "loss": 1.5669, "step": 5862 }, { "epoch": 0.6336323354587702, "grad_norm": 0.4921928346157074, "learning_rate": 8e-05, "loss": 2.0552, "step": 5863 }, { "epoch": 0.6337404085161569, "grad_norm": 0.3762911856174469, "learning_rate": 8e-05, "loss": 1.6095, "step": 5864 }, { "epoch": 0.6338484815735437, "grad_norm": 0.4185813367366791, "learning_rate": 8e-05, "loss": 1.6866, "step": 5865 }, { "epoch": 0.6339565546309305, "grad_norm": 0.42174768447875977, "learning_rate": 8e-05, "loss": 1.6516, "step": 5866 }, { "epoch": 0.6340646276883173, "grad_norm": 0.4447663426399231, "learning_rate": 8e-05, "loss": 1.7537, "step": 5867 }, { "epoch": 0.6341727007457041, "grad_norm": 0.47395235300064087, "learning_rate": 8e-05, "loss": 1.9245, "step": 5868 }, { "epoch": 0.6342807738030909, "grad_norm": 0.40187400579452515, "learning_rate": 8e-05, "loss": 1.6063, "step": 5869 }, { "epoch": 0.6343888468604777, "grad_norm": 0.4047797918319702, "learning_rate": 8e-05, "loss": 1.6298, "step": 5870 }, { "epoch": 0.6344969199178645, "grad_norm": 0.4176103174686432, "learning_rate": 8e-05, "loss": 1.7712, "step": 5871 }, { "epoch": 0.6346049929752513, "grad_norm": 0.40666237473487854, "learning_rate": 8e-05, "loss": 1.7215, "step": 5872 }, { "epoch": 0.6347130660326381, "grad_norm": 0.42609596252441406, "learning_rate": 8e-05, "loss": 1.6812, "step": 5873 }, { "epoch": 0.6348211390900249, "grad_norm": 0.38025882840156555, "learning_rate": 8e-05, "loss": 1.4906, "step": 5874 }, { "epoch": 0.6349292121474116, "grad_norm": 0.48950156569480896, "learning_rate": 8e-05, "loss": 1.742, "step": 5875 }, { "epoch": 0.6350372852047984, "grad_norm": 0.3881766200065613, "learning_rate": 8e-05, "loss": 1.4469, "step": 5876 }, { "epoch": 0.6351453582621852, "grad_norm": 0.4317636787891388, "learning_rate": 8e-05, "loss": 1.811, "step": 5877 }, { "epoch": 0.635253431319572, "grad_norm": 0.4312632381916046, "learning_rate": 8e-05, "loss": 1.7905, "step": 5878 }, { "epoch": 0.6353615043769588, "grad_norm": 0.4055800139904022, "learning_rate": 8e-05, "loss": 1.6172, "step": 5879 }, { "epoch": 0.6354695774343456, "grad_norm": 0.43827611207962036, "learning_rate": 8e-05, "loss": 1.7388, "step": 5880 }, { "epoch": 0.6355776504917324, "grad_norm": 0.4604509174823761, "learning_rate": 8e-05, "loss": 1.9248, "step": 5881 }, { "epoch": 0.6356857235491192, "grad_norm": 0.4025614857673645, "learning_rate": 8e-05, "loss": 1.3414, "step": 5882 }, { "epoch": 0.635793796606506, "grad_norm": 0.42184919118881226, "learning_rate": 8e-05, "loss": 1.6891, "step": 5883 }, { "epoch": 0.6359018696638928, "grad_norm": 0.41898685693740845, "learning_rate": 8e-05, "loss": 1.6021, "step": 5884 }, { "epoch": 0.6360099427212795, "grad_norm": 0.4092831015586853, "learning_rate": 8e-05, "loss": 1.8156, "step": 5885 }, { "epoch": 0.6361180157786663, "grad_norm": 0.44821423292160034, "learning_rate": 8e-05, "loss": 1.8157, "step": 5886 }, { "epoch": 0.6362260888360531, "grad_norm": 0.40425240993499756, "learning_rate": 8e-05, "loss": 1.4435, "step": 5887 }, { "epoch": 0.6363341618934399, "grad_norm": 0.43720507621765137, "learning_rate": 8e-05, "loss": 1.8235, "step": 5888 }, { "epoch": 0.6364422349508267, "grad_norm": 0.41527438163757324, "learning_rate": 8e-05, "loss": 1.6785, "step": 5889 }, { "epoch": 0.6365503080082136, "grad_norm": 0.41892772912979126, "learning_rate": 8e-05, "loss": 1.8348, "step": 5890 }, { "epoch": 0.6366583810656004, "grad_norm": 0.42598506808280945, "learning_rate": 8e-05, "loss": 1.7236, "step": 5891 }, { "epoch": 0.6367664541229872, "grad_norm": 0.43189412355422974, "learning_rate": 8e-05, "loss": 1.7066, "step": 5892 }, { "epoch": 0.636874527180374, "grad_norm": 0.3953350782394409, "learning_rate": 8e-05, "loss": 1.5864, "step": 5893 }, { "epoch": 0.6369826002377608, "grad_norm": 0.4174412190914154, "learning_rate": 8e-05, "loss": 1.5421, "step": 5894 }, { "epoch": 0.6370906732951476, "grad_norm": 0.42606857419013977, "learning_rate": 8e-05, "loss": 1.7985, "step": 5895 }, { "epoch": 0.6371987463525343, "grad_norm": 0.4193751811981201, "learning_rate": 8e-05, "loss": 1.8533, "step": 5896 }, { "epoch": 0.6373068194099211, "grad_norm": 0.42517632246017456, "learning_rate": 8e-05, "loss": 1.7113, "step": 5897 }, { "epoch": 0.6374148924673079, "grad_norm": 0.4157698154449463, "learning_rate": 8e-05, "loss": 1.6424, "step": 5898 }, { "epoch": 0.6375229655246947, "grad_norm": 0.42710694670677185, "learning_rate": 8e-05, "loss": 1.7425, "step": 5899 }, { "epoch": 0.6376310385820815, "grad_norm": 0.42149630188941956, "learning_rate": 8e-05, "loss": 1.6619, "step": 5900 }, { "epoch": 0.6377391116394683, "grad_norm": 0.4364214241504669, "learning_rate": 8e-05, "loss": 1.7178, "step": 5901 }, { "epoch": 0.6378471846968551, "grad_norm": 0.4101714491844177, "learning_rate": 8e-05, "loss": 1.6258, "step": 5902 }, { "epoch": 0.6379552577542419, "grad_norm": 0.4424304962158203, "learning_rate": 8e-05, "loss": 1.7939, "step": 5903 }, { "epoch": 0.6380633308116287, "grad_norm": 0.4274967610836029, "learning_rate": 8e-05, "loss": 1.8202, "step": 5904 }, { "epoch": 0.6381714038690155, "grad_norm": 0.41223785281181335, "learning_rate": 8e-05, "loss": 1.7155, "step": 5905 }, { "epoch": 0.6382794769264023, "grad_norm": 0.4205559194087982, "learning_rate": 8e-05, "loss": 1.7098, "step": 5906 }, { "epoch": 0.638387549983789, "grad_norm": 0.4731767177581787, "learning_rate": 8e-05, "loss": 1.7617, "step": 5907 }, { "epoch": 0.6384956230411758, "grad_norm": 0.41717997193336487, "learning_rate": 8e-05, "loss": 1.7957, "step": 5908 }, { "epoch": 0.6386036960985626, "grad_norm": 0.41988909244537354, "learning_rate": 8e-05, "loss": 1.7389, "step": 5909 }, { "epoch": 0.6387117691559494, "grad_norm": 0.4069858193397522, "learning_rate": 8e-05, "loss": 1.708, "step": 5910 }, { "epoch": 0.6388198422133362, "grad_norm": 0.3908928632736206, "learning_rate": 8e-05, "loss": 1.6975, "step": 5911 }, { "epoch": 0.638927915270723, "grad_norm": 0.4308492839336395, "learning_rate": 8e-05, "loss": 1.7535, "step": 5912 }, { "epoch": 0.6390359883281098, "grad_norm": 0.3947773873806, "learning_rate": 8e-05, "loss": 1.5992, "step": 5913 }, { "epoch": 0.6391440613854966, "grad_norm": 0.40337884426116943, "learning_rate": 8e-05, "loss": 1.6912, "step": 5914 }, { "epoch": 0.6392521344428834, "grad_norm": 0.44489115476608276, "learning_rate": 8e-05, "loss": 1.6515, "step": 5915 }, { "epoch": 0.6393602075002702, "grad_norm": 0.4458219110965729, "learning_rate": 8e-05, "loss": 1.5839, "step": 5916 }, { "epoch": 0.6394682805576569, "grad_norm": 0.5302431583404541, "learning_rate": 8e-05, "loss": 1.8071, "step": 5917 }, { "epoch": 0.6395763536150437, "grad_norm": 0.41243067383766174, "learning_rate": 8e-05, "loss": 1.5915, "step": 5918 }, { "epoch": 0.6396844266724305, "grad_norm": 0.441345751285553, "learning_rate": 8e-05, "loss": 1.7688, "step": 5919 }, { "epoch": 0.6397924997298173, "grad_norm": 0.4298408031463623, "learning_rate": 8e-05, "loss": 1.8117, "step": 5920 }, { "epoch": 0.6399005727872041, "grad_norm": 0.404812216758728, "learning_rate": 8e-05, "loss": 1.6704, "step": 5921 }, { "epoch": 0.6400086458445909, "grad_norm": 0.4513489902019501, "learning_rate": 8e-05, "loss": 1.8619, "step": 5922 }, { "epoch": 0.6401167189019777, "grad_norm": 0.4641247093677521, "learning_rate": 8e-05, "loss": 1.8823, "step": 5923 }, { "epoch": 0.6402247919593645, "grad_norm": 0.4099825918674469, "learning_rate": 8e-05, "loss": 1.729, "step": 5924 }, { "epoch": 0.6403328650167514, "grad_norm": 0.4017665684223175, "learning_rate": 8e-05, "loss": 1.6278, "step": 5925 }, { "epoch": 0.6404409380741382, "grad_norm": 0.5056819915771484, "learning_rate": 8e-05, "loss": 1.9833, "step": 5926 }, { "epoch": 0.640549011131525, "grad_norm": 0.4170728325843811, "learning_rate": 8e-05, "loss": 1.7419, "step": 5927 }, { "epoch": 0.6406570841889117, "grad_norm": 0.42615488171577454, "learning_rate": 8e-05, "loss": 1.8248, "step": 5928 }, { "epoch": 0.6407651572462985, "grad_norm": 0.4356894791126251, "learning_rate": 8e-05, "loss": 1.6913, "step": 5929 }, { "epoch": 0.6408732303036853, "grad_norm": 0.43304985761642456, "learning_rate": 8e-05, "loss": 1.7473, "step": 5930 }, { "epoch": 0.6409813033610721, "grad_norm": 0.40428903698921204, "learning_rate": 8e-05, "loss": 1.6087, "step": 5931 }, { "epoch": 0.6410893764184589, "grad_norm": 0.4069860875606537, "learning_rate": 8e-05, "loss": 1.5217, "step": 5932 }, { "epoch": 0.6411974494758457, "grad_norm": 0.4382326006889343, "learning_rate": 8e-05, "loss": 1.6056, "step": 5933 }, { "epoch": 0.6413055225332325, "grad_norm": 0.39903631806373596, "learning_rate": 8e-05, "loss": 1.5839, "step": 5934 }, { "epoch": 0.6414135955906193, "grad_norm": 0.4824443459510803, "learning_rate": 8e-05, "loss": 1.9891, "step": 5935 }, { "epoch": 0.6415216686480061, "grad_norm": 0.4243977665901184, "learning_rate": 8e-05, "loss": 1.5436, "step": 5936 }, { "epoch": 0.6416297417053929, "grad_norm": 0.4435882568359375, "learning_rate": 8e-05, "loss": 1.6656, "step": 5937 }, { "epoch": 0.6417378147627797, "grad_norm": 0.4456670880317688, "learning_rate": 8e-05, "loss": 1.839, "step": 5938 }, { "epoch": 0.6418458878201664, "grad_norm": 0.4523741602897644, "learning_rate": 8e-05, "loss": 1.6783, "step": 5939 }, { "epoch": 0.6419539608775532, "grad_norm": 0.4241046607494354, "learning_rate": 8e-05, "loss": 1.7067, "step": 5940 }, { "epoch": 0.64206203393494, "grad_norm": 0.40090230107307434, "learning_rate": 8e-05, "loss": 1.6194, "step": 5941 }, { "epoch": 0.6421701069923268, "grad_norm": 0.43487051129341125, "learning_rate": 8e-05, "loss": 1.8778, "step": 5942 }, { "epoch": 0.6422781800497136, "grad_norm": 0.37951064109802246, "learning_rate": 8e-05, "loss": 1.6027, "step": 5943 }, { "epoch": 0.6423862531071004, "grad_norm": 0.37684527039527893, "learning_rate": 8e-05, "loss": 1.5665, "step": 5944 }, { "epoch": 0.6424943261644872, "grad_norm": 0.4198205769062042, "learning_rate": 8e-05, "loss": 1.6908, "step": 5945 }, { "epoch": 0.642602399221874, "grad_norm": 0.4570189118385315, "learning_rate": 8e-05, "loss": 1.7282, "step": 5946 }, { "epoch": 0.6427104722792608, "grad_norm": 0.4548553228378296, "learning_rate": 8e-05, "loss": 1.6967, "step": 5947 }, { "epoch": 0.6428185453366476, "grad_norm": 0.4002516269683838, "learning_rate": 8e-05, "loss": 1.6702, "step": 5948 }, { "epoch": 0.6429266183940344, "grad_norm": 0.398049920797348, "learning_rate": 8e-05, "loss": 1.5799, "step": 5949 }, { "epoch": 0.6430346914514211, "grad_norm": 0.475448340177536, "learning_rate": 8e-05, "loss": 1.9621, "step": 5950 }, { "epoch": 0.6431427645088079, "grad_norm": 0.4206528067588806, "learning_rate": 8e-05, "loss": 1.7078, "step": 5951 }, { "epoch": 0.6432508375661947, "grad_norm": 0.41603797674179077, "learning_rate": 8e-05, "loss": 1.6493, "step": 5952 }, { "epoch": 0.6433589106235815, "grad_norm": 0.4449702799320221, "learning_rate": 8e-05, "loss": 1.6232, "step": 5953 }, { "epoch": 0.6434669836809683, "grad_norm": 0.3960975706577301, "learning_rate": 8e-05, "loss": 1.4676, "step": 5954 }, { "epoch": 0.6435750567383551, "grad_norm": 0.40631982684135437, "learning_rate": 8e-05, "loss": 1.7231, "step": 5955 }, { "epoch": 0.6436831297957419, "grad_norm": 0.4015238881111145, "learning_rate": 8e-05, "loss": 1.5872, "step": 5956 }, { "epoch": 0.6437912028531287, "grad_norm": 0.4078163504600525, "learning_rate": 8e-05, "loss": 1.6854, "step": 5957 }, { "epoch": 0.6438992759105155, "grad_norm": 0.4468514323234558, "learning_rate": 8e-05, "loss": 1.5877, "step": 5958 }, { "epoch": 0.6440073489679023, "grad_norm": 0.4127141535282135, "learning_rate": 8e-05, "loss": 1.7195, "step": 5959 }, { "epoch": 0.644115422025289, "grad_norm": 0.44148391485214233, "learning_rate": 8e-05, "loss": 1.7438, "step": 5960 }, { "epoch": 0.6442234950826758, "grad_norm": 0.5581531524658203, "learning_rate": 8e-05, "loss": 1.9885, "step": 5961 }, { "epoch": 0.6443315681400626, "grad_norm": 0.413822740316391, "learning_rate": 8e-05, "loss": 1.7774, "step": 5962 }, { "epoch": 0.6444396411974495, "grad_norm": 0.39981749653816223, "learning_rate": 8e-05, "loss": 1.6394, "step": 5963 }, { "epoch": 0.6445477142548363, "grad_norm": 0.39464741945266724, "learning_rate": 8e-05, "loss": 1.5967, "step": 5964 }, { "epoch": 0.6446557873122231, "grad_norm": 0.444970965385437, "learning_rate": 8e-05, "loss": 1.6563, "step": 5965 }, { "epoch": 0.6447638603696099, "grad_norm": 0.43060368299484253, "learning_rate": 8e-05, "loss": 1.6392, "step": 5966 }, { "epoch": 0.6448719334269967, "grad_norm": 0.40131300687789917, "learning_rate": 8e-05, "loss": 1.6946, "step": 5967 }, { "epoch": 0.6449800064843835, "grad_norm": 0.4246337413787842, "learning_rate": 8e-05, "loss": 1.6029, "step": 5968 }, { "epoch": 0.6450880795417703, "grad_norm": 0.3889176547527313, "learning_rate": 8e-05, "loss": 1.329, "step": 5969 }, { "epoch": 0.6451961525991571, "grad_norm": 0.4410610496997833, "learning_rate": 8e-05, "loss": 1.7425, "step": 5970 }, { "epoch": 0.6453042256565438, "grad_norm": 0.40063047409057617, "learning_rate": 8e-05, "loss": 1.4804, "step": 5971 }, { "epoch": 0.6454122987139306, "grad_norm": 0.4138284921646118, "learning_rate": 8e-05, "loss": 1.5908, "step": 5972 }, { "epoch": 0.6455203717713174, "grad_norm": 0.4406532347202301, "learning_rate": 8e-05, "loss": 1.7125, "step": 5973 }, { "epoch": 0.6456284448287042, "grad_norm": 0.4007008373737335, "learning_rate": 8e-05, "loss": 1.6404, "step": 5974 }, { "epoch": 0.645736517886091, "grad_norm": 0.462288498878479, "learning_rate": 8e-05, "loss": 1.778, "step": 5975 }, { "epoch": 0.6458445909434778, "grad_norm": 0.40858885645866394, "learning_rate": 8e-05, "loss": 1.5464, "step": 5976 }, { "epoch": 0.6459526640008646, "grad_norm": 0.4496130347251892, "learning_rate": 8e-05, "loss": 1.9094, "step": 5977 }, { "epoch": 0.6460607370582514, "grad_norm": 0.48654109239578247, "learning_rate": 8e-05, "loss": 1.712, "step": 5978 }, { "epoch": 0.6461688101156382, "grad_norm": 0.4343802332878113, "learning_rate": 8e-05, "loss": 1.5099, "step": 5979 }, { "epoch": 0.646276883173025, "grad_norm": 0.4566666781902313, "learning_rate": 8e-05, "loss": 1.808, "step": 5980 }, { "epoch": 0.6463849562304118, "grad_norm": 0.46233969926834106, "learning_rate": 8e-05, "loss": 1.8845, "step": 5981 }, { "epoch": 0.6464930292877985, "grad_norm": 0.49122631549835205, "learning_rate": 8e-05, "loss": 1.8415, "step": 5982 }, { "epoch": 0.6466011023451853, "grad_norm": 0.4473550617694855, "learning_rate": 8e-05, "loss": 1.9222, "step": 5983 }, { "epoch": 0.6467091754025721, "grad_norm": 0.42812246084213257, "learning_rate": 8e-05, "loss": 1.7767, "step": 5984 }, { "epoch": 0.6468172484599589, "grad_norm": 0.41955262422561646, "learning_rate": 8e-05, "loss": 1.7288, "step": 5985 }, { "epoch": 0.6469253215173457, "grad_norm": 0.5279108285903931, "learning_rate": 8e-05, "loss": 1.8856, "step": 5986 }, { "epoch": 0.6470333945747325, "grad_norm": 0.440212219953537, "learning_rate": 8e-05, "loss": 1.8732, "step": 5987 }, { "epoch": 0.6471414676321193, "grad_norm": 0.4515148401260376, "learning_rate": 8e-05, "loss": 1.7323, "step": 5988 }, { "epoch": 0.6472495406895061, "grad_norm": 0.46701234579086304, "learning_rate": 8e-05, "loss": 1.7868, "step": 5989 }, { "epoch": 0.6473576137468929, "grad_norm": 0.4646361172199249, "learning_rate": 8e-05, "loss": 1.7771, "step": 5990 }, { "epoch": 0.6474656868042797, "grad_norm": 0.4157301187515259, "learning_rate": 8e-05, "loss": 1.6421, "step": 5991 }, { "epoch": 0.6475737598616664, "grad_norm": 0.4450364410877228, "learning_rate": 8e-05, "loss": 1.6027, "step": 5992 }, { "epoch": 0.6476818329190532, "grad_norm": 0.4636766016483307, "learning_rate": 8e-05, "loss": 1.7481, "step": 5993 }, { "epoch": 0.64778990597644, "grad_norm": 0.4249369502067566, "learning_rate": 8e-05, "loss": 1.6527, "step": 5994 }, { "epoch": 0.6478979790338268, "grad_norm": 0.41753992438316345, "learning_rate": 8e-05, "loss": 1.7086, "step": 5995 }, { "epoch": 0.6480060520912136, "grad_norm": 0.4141181409358978, "learning_rate": 8e-05, "loss": 1.7583, "step": 5996 }, { "epoch": 0.6481141251486005, "grad_norm": 0.4067273736000061, "learning_rate": 8e-05, "loss": 1.4903, "step": 5997 }, { "epoch": 0.6482221982059873, "grad_norm": 0.44347625970840454, "learning_rate": 8e-05, "loss": 1.4582, "step": 5998 }, { "epoch": 0.6483302712633741, "grad_norm": 0.40376442670822144, "learning_rate": 8e-05, "loss": 1.6974, "step": 5999 }, { "epoch": 0.6484383443207609, "grad_norm": 0.4294092357158661, "learning_rate": 8e-05, "loss": 1.6185, "step": 6000 }, { "epoch": 0.6485464173781477, "grad_norm": 0.41296014189720154, "learning_rate": 8e-05, "loss": 1.5639, "step": 6001 }, { "epoch": 0.6486544904355345, "grad_norm": 0.4245632290840149, "learning_rate": 8e-05, "loss": 1.7482, "step": 6002 }, { "epoch": 0.6487625634929212, "grad_norm": 0.4372320771217346, "learning_rate": 8e-05, "loss": 1.6191, "step": 6003 }, { "epoch": 0.648870636550308, "grad_norm": 0.4616338610649109, "learning_rate": 8e-05, "loss": 1.6431, "step": 6004 }, { "epoch": 0.6489787096076948, "grad_norm": 0.41995561122894287, "learning_rate": 8e-05, "loss": 1.6753, "step": 6005 }, { "epoch": 0.6490867826650816, "grad_norm": 0.4097039997577667, "learning_rate": 8e-05, "loss": 1.6599, "step": 6006 }, { "epoch": 0.6491948557224684, "grad_norm": 0.42022982239723206, "learning_rate": 8e-05, "loss": 1.6784, "step": 6007 }, { "epoch": 0.6493029287798552, "grad_norm": 0.43660926818847656, "learning_rate": 8e-05, "loss": 1.7272, "step": 6008 }, { "epoch": 0.649411001837242, "grad_norm": 0.4854525923728943, "learning_rate": 8e-05, "loss": 1.6608, "step": 6009 }, { "epoch": 0.6495190748946288, "grad_norm": 0.41183900833129883, "learning_rate": 8e-05, "loss": 1.7181, "step": 6010 }, { "epoch": 0.6496271479520156, "grad_norm": 0.40540412068367004, "learning_rate": 8e-05, "loss": 1.728, "step": 6011 }, { "epoch": 0.6497352210094024, "grad_norm": 0.4309687912464142, "learning_rate": 8e-05, "loss": 1.7552, "step": 6012 }, { "epoch": 0.6498432940667892, "grad_norm": 0.42914077639579773, "learning_rate": 8e-05, "loss": 1.6169, "step": 6013 }, { "epoch": 0.6499513671241759, "grad_norm": 0.4755520522594452, "learning_rate": 8e-05, "loss": 1.9241, "step": 6014 }, { "epoch": 0.6500594401815627, "grad_norm": 0.39908960461616516, "learning_rate": 8e-05, "loss": 1.8235, "step": 6015 }, { "epoch": 0.6501675132389495, "grad_norm": 0.40657928586006165, "learning_rate": 8e-05, "loss": 1.6772, "step": 6016 }, { "epoch": 0.6502755862963363, "grad_norm": 0.4402332901954651, "learning_rate": 8e-05, "loss": 1.4134, "step": 6017 }, { "epoch": 0.6503836593537231, "grad_norm": 0.41817325353622437, "learning_rate": 8e-05, "loss": 1.7617, "step": 6018 }, { "epoch": 0.6504917324111099, "grad_norm": 0.42117205262184143, "learning_rate": 8e-05, "loss": 1.6436, "step": 6019 }, { "epoch": 0.6505998054684967, "grad_norm": 0.4075367748737335, "learning_rate": 8e-05, "loss": 1.519, "step": 6020 }, { "epoch": 0.6507078785258835, "grad_norm": 0.3931729793548584, "learning_rate": 8e-05, "loss": 1.6352, "step": 6021 }, { "epoch": 0.6508159515832703, "grad_norm": 0.4228362739086151, "learning_rate": 8e-05, "loss": 1.6317, "step": 6022 }, { "epoch": 0.6509240246406571, "grad_norm": 0.4444679021835327, "learning_rate": 8e-05, "loss": 1.8934, "step": 6023 }, { "epoch": 0.6510320976980439, "grad_norm": 0.44917336106300354, "learning_rate": 8e-05, "loss": 1.8571, "step": 6024 }, { "epoch": 0.6511401707554306, "grad_norm": 0.4502103328704834, "learning_rate": 8e-05, "loss": 1.812, "step": 6025 }, { "epoch": 0.6512482438128174, "grad_norm": 0.4338429868221283, "learning_rate": 8e-05, "loss": 1.8018, "step": 6026 }, { "epoch": 0.6513563168702042, "grad_norm": 0.42189955711364746, "learning_rate": 8e-05, "loss": 1.7334, "step": 6027 }, { "epoch": 0.651464389927591, "grad_norm": 0.45813271403312683, "learning_rate": 8e-05, "loss": 1.9343, "step": 6028 }, { "epoch": 0.6515724629849778, "grad_norm": 0.41653189063072205, "learning_rate": 8e-05, "loss": 1.7613, "step": 6029 }, { "epoch": 0.6516805360423646, "grad_norm": 0.39771711826324463, "learning_rate": 8e-05, "loss": 1.7, "step": 6030 }, { "epoch": 0.6517886090997514, "grad_norm": 0.41165611147880554, "learning_rate": 8e-05, "loss": 1.6719, "step": 6031 }, { "epoch": 0.6518966821571383, "grad_norm": 0.41300857067108154, "learning_rate": 8e-05, "loss": 1.6307, "step": 6032 }, { "epoch": 0.6520047552145251, "grad_norm": 0.44221505522727966, "learning_rate": 8e-05, "loss": 1.6646, "step": 6033 }, { "epoch": 0.6521128282719119, "grad_norm": 0.3943091332912445, "learning_rate": 8e-05, "loss": 1.5554, "step": 6034 }, { "epoch": 0.6522209013292986, "grad_norm": 0.4221917688846588, "learning_rate": 8e-05, "loss": 1.5843, "step": 6035 }, { "epoch": 0.6523289743866854, "grad_norm": 0.4098266065120697, "learning_rate": 8e-05, "loss": 1.6481, "step": 6036 }, { "epoch": 0.6524370474440722, "grad_norm": 0.45125260949134827, "learning_rate": 8e-05, "loss": 1.8059, "step": 6037 }, { "epoch": 0.652545120501459, "grad_norm": 0.40337345004081726, "learning_rate": 8e-05, "loss": 1.8275, "step": 6038 }, { "epoch": 0.6526531935588458, "grad_norm": 0.4045194983482361, "learning_rate": 8e-05, "loss": 1.6181, "step": 6039 }, { "epoch": 0.6527612666162326, "grad_norm": 0.4071527123451233, "learning_rate": 8e-05, "loss": 1.3589, "step": 6040 }, { "epoch": 0.6528693396736194, "grad_norm": 0.40541642904281616, "learning_rate": 8e-05, "loss": 1.8038, "step": 6041 }, { "epoch": 0.6529774127310062, "grad_norm": 0.4687822461128235, "learning_rate": 8e-05, "loss": 1.8491, "step": 6042 }, { "epoch": 0.653085485788393, "grad_norm": 0.4163665473461151, "learning_rate": 8e-05, "loss": 1.7271, "step": 6043 }, { "epoch": 0.6531935588457798, "grad_norm": 0.49661147594451904, "learning_rate": 8e-05, "loss": 1.7281, "step": 6044 }, { "epoch": 0.6533016319031666, "grad_norm": 0.3979349434375763, "learning_rate": 8e-05, "loss": 1.447, "step": 6045 }, { "epoch": 0.6534097049605533, "grad_norm": 0.4367724359035492, "learning_rate": 8e-05, "loss": 1.8408, "step": 6046 }, { "epoch": 0.6535177780179401, "grad_norm": 0.45611637830734253, "learning_rate": 8e-05, "loss": 1.8178, "step": 6047 }, { "epoch": 0.6536258510753269, "grad_norm": 0.4215480387210846, "learning_rate": 8e-05, "loss": 1.591, "step": 6048 }, { "epoch": 0.6537339241327137, "grad_norm": 0.4396744668483734, "learning_rate": 8e-05, "loss": 1.6705, "step": 6049 }, { "epoch": 0.6538419971901005, "grad_norm": 0.4155380427837372, "learning_rate": 8e-05, "loss": 1.3962, "step": 6050 }, { "epoch": 0.6539500702474873, "grad_norm": 0.42172956466674805, "learning_rate": 8e-05, "loss": 1.5569, "step": 6051 }, { "epoch": 0.6540581433048741, "grad_norm": 0.47426798939704895, "learning_rate": 8e-05, "loss": 1.6266, "step": 6052 }, { "epoch": 0.6541662163622609, "grad_norm": 0.41053324937820435, "learning_rate": 8e-05, "loss": 1.6672, "step": 6053 }, { "epoch": 0.6542742894196477, "grad_norm": 0.433466374874115, "learning_rate": 8e-05, "loss": 1.7865, "step": 6054 }, { "epoch": 0.6543823624770345, "grad_norm": 0.4832453429698944, "learning_rate": 8e-05, "loss": 1.608, "step": 6055 }, { "epoch": 0.6544904355344213, "grad_norm": 0.43349623680114746, "learning_rate": 8e-05, "loss": 1.7102, "step": 6056 }, { "epoch": 0.654598508591808, "grad_norm": 0.430059015750885, "learning_rate": 8e-05, "loss": 1.5781, "step": 6057 }, { "epoch": 0.6547065816491948, "grad_norm": 0.4353489875793457, "learning_rate": 8e-05, "loss": 1.6483, "step": 6058 }, { "epoch": 0.6548146547065816, "grad_norm": 0.4279748499393463, "learning_rate": 8e-05, "loss": 1.7894, "step": 6059 }, { "epoch": 0.6549227277639684, "grad_norm": 0.4506438076496124, "learning_rate": 8e-05, "loss": 1.594, "step": 6060 }, { "epoch": 0.6550308008213552, "grad_norm": 0.42169657349586487, "learning_rate": 8e-05, "loss": 1.6413, "step": 6061 }, { "epoch": 0.655138873878742, "grad_norm": 0.40544694662094116, "learning_rate": 8e-05, "loss": 1.5869, "step": 6062 }, { "epoch": 0.6552469469361288, "grad_norm": 0.40419498085975647, "learning_rate": 8e-05, "loss": 1.6336, "step": 6063 }, { "epoch": 0.6553550199935156, "grad_norm": 0.43076175451278687, "learning_rate": 8e-05, "loss": 1.6936, "step": 6064 }, { "epoch": 0.6554630930509024, "grad_norm": 0.44776099920272827, "learning_rate": 8e-05, "loss": 1.8426, "step": 6065 }, { "epoch": 0.6555711661082892, "grad_norm": 0.4287852346897125, "learning_rate": 8e-05, "loss": 1.836, "step": 6066 }, { "epoch": 0.6556792391656759, "grad_norm": 0.4299076795578003, "learning_rate": 8e-05, "loss": 1.6631, "step": 6067 }, { "epoch": 0.6557873122230627, "grad_norm": 0.5212808847427368, "learning_rate": 8e-05, "loss": 1.9388, "step": 6068 }, { "epoch": 0.6558953852804495, "grad_norm": 0.4532490670681, "learning_rate": 8e-05, "loss": 1.7447, "step": 6069 }, { "epoch": 0.6560034583378364, "grad_norm": 0.4737337827682495, "learning_rate": 8e-05, "loss": 1.9207, "step": 6070 }, { "epoch": 0.6561115313952232, "grad_norm": 0.3897891342639923, "learning_rate": 8e-05, "loss": 1.597, "step": 6071 }, { "epoch": 0.65621960445261, "grad_norm": 0.4140532314777374, "learning_rate": 8e-05, "loss": 1.6137, "step": 6072 }, { "epoch": 0.6563276775099968, "grad_norm": 0.440957635641098, "learning_rate": 8e-05, "loss": 1.7261, "step": 6073 }, { "epoch": 0.6564357505673836, "grad_norm": 0.40956977009773254, "learning_rate": 8e-05, "loss": 1.6031, "step": 6074 }, { "epoch": 0.6565438236247704, "grad_norm": 0.43229418992996216, "learning_rate": 8e-05, "loss": 1.6924, "step": 6075 }, { "epoch": 0.6566518966821572, "grad_norm": 0.43051499128341675, "learning_rate": 8e-05, "loss": 1.7556, "step": 6076 }, { "epoch": 0.656759969739544, "grad_norm": 0.40451645851135254, "learning_rate": 8e-05, "loss": 1.6311, "step": 6077 }, { "epoch": 0.6568680427969307, "grad_norm": 0.4123532176017761, "learning_rate": 8e-05, "loss": 1.601, "step": 6078 }, { "epoch": 0.6569761158543175, "grad_norm": 0.3901147246360779, "learning_rate": 8e-05, "loss": 1.5906, "step": 6079 }, { "epoch": 0.6570841889117043, "grad_norm": 0.4197995662689209, "learning_rate": 8e-05, "loss": 1.6575, "step": 6080 }, { "epoch": 0.6571922619690911, "grad_norm": 0.4124607443809509, "learning_rate": 8e-05, "loss": 1.6005, "step": 6081 }, { "epoch": 0.6573003350264779, "grad_norm": 0.401763916015625, "learning_rate": 8e-05, "loss": 1.6691, "step": 6082 }, { "epoch": 0.6574084080838647, "grad_norm": 0.4340497851371765, "learning_rate": 8e-05, "loss": 1.6414, "step": 6083 }, { "epoch": 0.6575164811412515, "grad_norm": 0.40181589126586914, "learning_rate": 8e-05, "loss": 1.6396, "step": 6084 }, { "epoch": 0.6576245541986383, "grad_norm": 0.4486393928527832, "learning_rate": 8e-05, "loss": 1.7519, "step": 6085 }, { "epoch": 0.6577326272560251, "grad_norm": 0.4246681332588196, "learning_rate": 8e-05, "loss": 1.533, "step": 6086 }, { "epoch": 0.6578407003134119, "grad_norm": 0.4050711393356323, "learning_rate": 8e-05, "loss": 1.5923, "step": 6087 }, { "epoch": 0.6579487733707987, "grad_norm": 0.46707841753959656, "learning_rate": 8e-05, "loss": 1.7225, "step": 6088 }, { "epoch": 0.6580568464281854, "grad_norm": 0.5020071864128113, "learning_rate": 8e-05, "loss": 1.7475, "step": 6089 }, { "epoch": 0.6581649194855722, "grad_norm": 0.4412035644054413, "learning_rate": 8e-05, "loss": 1.7253, "step": 6090 }, { "epoch": 0.658272992542959, "grad_norm": 0.4276970624923706, "learning_rate": 8e-05, "loss": 1.5957, "step": 6091 }, { "epoch": 0.6583810656003458, "grad_norm": 0.40917789936065674, "learning_rate": 8e-05, "loss": 1.5658, "step": 6092 }, { "epoch": 0.6584891386577326, "grad_norm": 0.43150508403778076, "learning_rate": 8e-05, "loss": 1.6093, "step": 6093 }, { "epoch": 0.6585972117151194, "grad_norm": 0.4382350742816925, "learning_rate": 8e-05, "loss": 1.8298, "step": 6094 }, { "epoch": 0.6587052847725062, "grad_norm": 0.4191981256008148, "learning_rate": 8e-05, "loss": 1.714, "step": 6095 }, { "epoch": 0.658813357829893, "grad_norm": 0.42044052481651306, "learning_rate": 8e-05, "loss": 1.722, "step": 6096 }, { "epoch": 0.6589214308872798, "grad_norm": 0.38723838329315186, "learning_rate": 8e-05, "loss": 1.5046, "step": 6097 }, { "epoch": 0.6590295039446666, "grad_norm": 0.48346540331840515, "learning_rate": 8e-05, "loss": 1.8258, "step": 6098 }, { "epoch": 0.6591375770020534, "grad_norm": 0.39127472043037415, "learning_rate": 8e-05, "loss": 1.5595, "step": 6099 }, { "epoch": 0.6592456500594401, "grad_norm": 0.40537029504776, "learning_rate": 8e-05, "loss": 1.6142, "step": 6100 }, { "epoch": 0.6593537231168269, "grad_norm": 0.42720553278923035, "learning_rate": 8e-05, "loss": 1.7284, "step": 6101 }, { "epoch": 0.6594617961742137, "grad_norm": 0.4623187780380249, "learning_rate": 8e-05, "loss": 1.9205, "step": 6102 }, { "epoch": 0.6595698692316005, "grad_norm": 0.3741806745529175, "learning_rate": 8e-05, "loss": 1.627, "step": 6103 }, { "epoch": 0.6596779422889874, "grad_norm": 0.4184570014476776, "learning_rate": 8e-05, "loss": 1.5865, "step": 6104 }, { "epoch": 0.6597860153463742, "grad_norm": 0.42817896604537964, "learning_rate": 8e-05, "loss": 1.7417, "step": 6105 }, { "epoch": 0.659894088403761, "grad_norm": 0.395384818315506, "learning_rate": 8e-05, "loss": 1.5811, "step": 6106 }, { "epoch": 0.6600021614611478, "grad_norm": 0.46312108635902405, "learning_rate": 8e-05, "loss": 1.936, "step": 6107 }, { "epoch": 0.6601102345185346, "grad_norm": 0.4472822844982147, "learning_rate": 8e-05, "loss": 1.8055, "step": 6108 }, { "epoch": 0.6602183075759214, "grad_norm": 0.4232169985771179, "learning_rate": 8e-05, "loss": 1.6387, "step": 6109 }, { "epoch": 0.6603263806333081, "grad_norm": 0.44168177247047424, "learning_rate": 8e-05, "loss": 1.7572, "step": 6110 }, { "epoch": 0.6604344536906949, "grad_norm": 0.4148571491241455, "learning_rate": 8e-05, "loss": 1.6532, "step": 6111 }, { "epoch": 0.6605425267480817, "grad_norm": 0.41171786189079285, "learning_rate": 8e-05, "loss": 1.6283, "step": 6112 }, { "epoch": 0.6606505998054685, "grad_norm": 0.4125647246837616, "learning_rate": 8e-05, "loss": 1.6605, "step": 6113 }, { "epoch": 0.6607586728628553, "grad_norm": 0.42169350385665894, "learning_rate": 8e-05, "loss": 1.4374, "step": 6114 }, { "epoch": 0.6608667459202421, "grad_norm": 0.4501084089279175, "learning_rate": 8e-05, "loss": 1.7529, "step": 6115 }, { "epoch": 0.6609748189776289, "grad_norm": 0.4003376364707947, "learning_rate": 8e-05, "loss": 1.601, "step": 6116 }, { "epoch": 0.6610828920350157, "grad_norm": 0.45977428555488586, "learning_rate": 8e-05, "loss": 1.8334, "step": 6117 }, { "epoch": 0.6611909650924025, "grad_norm": 0.46009179949760437, "learning_rate": 8e-05, "loss": 1.4361, "step": 6118 }, { "epoch": 0.6612990381497893, "grad_norm": 0.43959900736808777, "learning_rate": 8e-05, "loss": 1.7649, "step": 6119 }, { "epoch": 0.6614071112071761, "grad_norm": 0.469620019197464, "learning_rate": 8e-05, "loss": 1.7539, "step": 6120 }, { "epoch": 0.6615151842645628, "grad_norm": 0.4284898042678833, "learning_rate": 8e-05, "loss": 1.7029, "step": 6121 }, { "epoch": 0.6616232573219496, "grad_norm": 0.4549995958805084, "learning_rate": 8e-05, "loss": 1.6792, "step": 6122 }, { "epoch": 0.6617313303793364, "grad_norm": 0.4103977084159851, "learning_rate": 8e-05, "loss": 1.6468, "step": 6123 }, { "epoch": 0.6618394034367232, "grad_norm": 0.4316926896572113, "learning_rate": 8e-05, "loss": 1.6111, "step": 6124 }, { "epoch": 0.66194747649411, "grad_norm": 0.4194194972515106, "learning_rate": 8e-05, "loss": 1.7083, "step": 6125 }, { "epoch": 0.6620555495514968, "grad_norm": 0.4177124798297882, "learning_rate": 8e-05, "loss": 1.8269, "step": 6126 }, { "epoch": 0.6621636226088836, "grad_norm": 0.40558698773384094, "learning_rate": 8e-05, "loss": 1.6732, "step": 6127 }, { "epoch": 0.6622716956662704, "grad_norm": 0.4189439117908478, "learning_rate": 8e-05, "loss": 1.6792, "step": 6128 }, { "epoch": 0.6623797687236572, "grad_norm": 0.421906977891922, "learning_rate": 8e-05, "loss": 1.8913, "step": 6129 }, { "epoch": 0.662487841781044, "grad_norm": 0.41307875514030457, "learning_rate": 8e-05, "loss": 1.637, "step": 6130 }, { "epoch": 0.6625959148384308, "grad_norm": 0.42005041241645813, "learning_rate": 8e-05, "loss": 1.7172, "step": 6131 }, { "epoch": 0.6627039878958175, "grad_norm": 0.450188010931015, "learning_rate": 8e-05, "loss": 1.7871, "step": 6132 }, { "epoch": 0.6628120609532043, "grad_norm": 0.43893492221832275, "learning_rate": 8e-05, "loss": 1.7452, "step": 6133 }, { "epoch": 0.6629201340105911, "grad_norm": 0.44087547063827515, "learning_rate": 8e-05, "loss": 1.8274, "step": 6134 }, { "epoch": 0.6630282070679779, "grad_norm": 0.42931675910949707, "learning_rate": 8e-05, "loss": 1.8968, "step": 6135 }, { "epoch": 0.6631362801253647, "grad_norm": 0.42867156863212585, "learning_rate": 8e-05, "loss": 1.745, "step": 6136 }, { "epoch": 0.6632443531827515, "grad_norm": 0.4360094666481018, "learning_rate": 8e-05, "loss": 1.6567, "step": 6137 }, { "epoch": 0.6633524262401383, "grad_norm": 0.3986239731311798, "learning_rate": 8e-05, "loss": 1.4479, "step": 6138 }, { "epoch": 0.6634604992975252, "grad_norm": 0.4227740168571472, "learning_rate": 8e-05, "loss": 1.5803, "step": 6139 }, { "epoch": 0.663568572354912, "grad_norm": 0.43978676199913025, "learning_rate": 8e-05, "loss": 1.7664, "step": 6140 }, { "epoch": 0.6636766454122988, "grad_norm": 0.45683377981185913, "learning_rate": 8e-05, "loss": 1.7261, "step": 6141 }, { "epoch": 0.6637847184696855, "grad_norm": 0.40478983521461487, "learning_rate": 8e-05, "loss": 1.7067, "step": 6142 }, { "epoch": 0.6638927915270723, "grad_norm": 0.41596484184265137, "learning_rate": 8e-05, "loss": 1.8087, "step": 6143 }, { "epoch": 0.6640008645844591, "grad_norm": 0.42229163646698, "learning_rate": 8e-05, "loss": 1.5987, "step": 6144 }, { "epoch": 0.6641089376418459, "grad_norm": 0.4067520797252655, "learning_rate": 8e-05, "loss": 1.6915, "step": 6145 }, { "epoch": 0.6642170106992327, "grad_norm": 0.4279630482196808, "learning_rate": 8e-05, "loss": 1.5516, "step": 6146 }, { "epoch": 0.6643250837566195, "grad_norm": 0.37149715423583984, "learning_rate": 8e-05, "loss": 1.5021, "step": 6147 }, { "epoch": 0.6644331568140063, "grad_norm": 0.48137471079826355, "learning_rate": 8e-05, "loss": 1.7254, "step": 6148 }, { "epoch": 0.6645412298713931, "grad_norm": 0.4220965504646301, "learning_rate": 8e-05, "loss": 1.5903, "step": 6149 }, { "epoch": 0.6646493029287799, "grad_norm": 0.46038171648979187, "learning_rate": 8e-05, "loss": 1.7462, "step": 6150 }, { "epoch": 0.6647573759861667, "grad_norm": 0.45745348930358887, "learning_rate": 8e-05, "loss": 1.7477, "step": 6151 }, { "epoch": 0.6648654490435535, "grad_norm": 0.47161728143692017, "learning_rate": 8e-05, "loss": 1.7438, "step": 6152 }, { "epoch": 0.6649735221009402, "grad_norm": 0.45101794600486755, "learning_rate": 8e-05, "loss": 1.7504, "step": 6153 }, { "epoch": 0.665081595158327, "grad_norm": 0.4284598231315613, "learning_rate": 8e-05, "loss": 1.7772, "step": 6154 }, { "epoch": 0.6651896682157138, "grad_norm": 0.4226823151111603, "learning_rate": 8e-05, "loss": 1.7547, "step": 6155 }, { "epoch": 0.6652977412731006, "grad_norm": 0.39508283138275146, "learning_rate": 8e-05, "loss": 1.5702, "step": 6156 }, { "epoch": 0.6654058143304874, "grad_norm": 0.4131785035133362, "learning_rate": 8e-05, "loss": 1.748, "step": 6157 }, { "epoch": 0.6655138873878742, "grad_norm": 0.4220304489135742, "learning_rate": 8e-05, "loss": 1.7057, "step": 6158 }, { "epoch": 0.665621960445261, "grad_norm": 0.3927169442176819, "learning_rate": 8e-05, "loss": 1.4676, "step": 6159 }, { "epoch": 0.6657300335026478, "grad_norm": 0.48370683193206787, "learning_rate": 8e-05, "loss": 1.7351, "step": 6160 }, { "epoch": 0.6658381065600346, "grad_norm": 0.4398593604564667, "learning_rate": 8e-05, "loss": 1.5907, "step": 6161 }, { "epoch": 0.6659461796174214, "grad_norm": 0.4521885812282562, "learning_rate": 8e-05, "loss": 1.6259, "step": 6162 }, { "epoch": 0.6660542526748082, "grad_norm": 0.402980238199234, "learning_rate": 8e-05, "loss": 1.5847, "step": 6163 }, { "epoch": 0.6661623257321949, "grad_norm": 0.4150359630584717, "learning_rate": 8e-05, "loss": 1.745, "step": 6164 }, { "epoch": 0.6662703987895817, "grad_norm": 0.4241977632045746, "learning_rate": 8e-05, "loss": 1.5603, "step": 6165 }, { "epoch": 0.6663784718469685, "grad_norm": 0.3871896266937256, "learning_rate": 8e-05, "loss": 1.5511, "step": 6166 }, { "epoch": 0.6664865449043553, "grad_norm": 0.45419231057167053, "learning_rate": 8e-05, "loss": 1.9366, "step": 6167 }, { "epoch": 0.6665946179617421, "grad_norm": 0.4616711437702179, "learning_rate": 8e-05, "loss": 1.8016, "step": 6168 }, { "epoch": 0.6667026910191289, "grad_norm": 0.4245321452617645, "learning_rate": 8e-05, "loss": 1.6511, "step": 6169 }, { "epoch": 0.6668107640765157, "grad_norm": 0.4353701174259186, "learning_rate": 8e-05, "loss": 1.6868, "step": 6170 }, { "epoch": 0.6669188371339025, "grad_norm": 0.42061108350753784, "learning_rate": 8e-05, "loss": 1.6658, "step": 6171 }, { "epoch": 0.6670269101912893, "grad_norm": 0.4456539750099182, "learning_rate": 8e-05, "loss": 1.7678, "step": 6172 }, { "epoch": 0.6671349832486761, "grad_norm": 0.4583442807197571, "learning_rate": 8e-05, "loss": 1.7124, "step": 6173 }, { "epoch": 0.667243056306063, "grad_norm": 0.41840746998786926, "learning_rate": 8e-05, "loss": 1.7058, "step": 6174 }, { "epoch": 0.6673511293634496, "grad_norm": 0.43913835287094116, "learning_rate": 8e-05, "loss": 1.8152, "step": 6175 }, { "epoch": 0.6674592024208365, "grad_norm": 0.3967171013355255, "learning_rate": 8e-05, "loss": 1.6801, "step": 6176 }, { "epoch": 0.6675672754782233, "grad_norm": 0.4275089502334595, "learning_rate": 8e-05, "loss": 1.7505, "step": 6177 }, { "epoch": 0.6676753485356101, "grad_norm": 0.41466858983039856, "learning_rate": 8e-05, "loss": 1.5788, "step": 6178 }, { "epoch": 0.6677834215929969, "grad_norm": 0.48071107268333435, "learning_rate": 8e-05, "loss": 1.778, "step": 6179 }, { "epoch": 0.6678914946503837, "grad_norm": 0.43582066893577576, "learning_rate": 8e-05, "loss": 1.7655, "step": 6180 }, { "epoch": 0.6679995677077705, "grad_norm": 0.4705914258956909, "learning_rate": 8e-05, "loss": 1.698, "step": 6181 }, { "epoch": 0.6681076407651573, "grad_norm": 0.47086799144744873, "learning_rate": 8e-05, "loss": 1.9766, "step": 6182 }, { "epoch": 0.6682157138225441, "grad_norm": 0.42524734139442444, "learning_rate": 8e-05, "loss": 1.7035, "step": 6183 }, { "epoch": 0.6683237868799309, "grad_norm": 0.4200817346572876, "learning_rate": 8e-05, "loss": 1.7446, "step": 6184 }, { "epoch": 0.6684318599373176, "grad_norm": 0.42611342668533325, "learning_rate": 8e-05, "loss": 1.5458, "step": 6185 }, { "epoch": 0.6685399329947044, "grad_norm": 0.4654000997543335, "learning_rate": 8e-05, "loss": 1.8695, "step": 6186 }, { "epoch": 0.6686480060520912, "grad_norm": 0.42026063799858093, "learning_rate": 8e-05, "loss": 1.5133, "step": 6187 }, { "epoch": 0.668756079109478, "grad_norm": 0.42000868916511536, "learning_rate": 8e-05, "loss": 1.6931, "step": 6188 }, { "epoch": 0.6688641521668648, "grad_norm": 0.42054906487464905, "learning_rate": 8e-05, "loss": 1.7435, "step": 6189 }, { "epoch": 0.6689722252242516, "grad_norm": 0.4656379520893097, "learning_rate": 8e-05, "loss": 1.6984, "step": 6190 }, { "epoch": 0.6690802982816384, "grad_norm": 0.4375896453857422, "learning_rate": 8e-05, "loss": 1.7733, "step": 6191 }, { "epoch": 0.6691883713390252, "grad_norm": 0.4236452877521515, "learning_rate": 8e-05, "loss": 1.8175, "step": 6192 }, { "epoch": 0.669296444396412, "grad_norm": 0.42429712414741516, "learning_rate": 8e-05, "loss": 1.612, "step": 6193 }, { "epoch": 0.6694045174537988, "grad_norm": 0.4229009747505188, "learning_rate": 8e-05, "loss": 1.6819, "step": 6194 }, { "epoch": 0.6695125905111856, "grad_norm": 0.451811820268631, "learning_rate": 8e-05, "loss": 1.7632, "step": 6195 }, { "epoch": 0.6696206635685723, "grad_norm": 0.46375322341918945, "learning_rate": 8e-05, "loss": 1.7986, "step": 6196 }, { "epoch": 0.6697287366259591, "grad_norm": 0.4495725929737091, "learning_rate": 8e-05, "loss": 1.6764, "step": 6197 }, { "epoch": 0.6698368096833459, "grad_norm": 0.4135211706161499, "learning_rate": 8e-05, "loss": 1.6079, "step": 6198 }, { "epoch": 0.6699448827407327, "grad_norm": 0.4611162841320038, "learning_rate": 8e-05, "loss": 1.6928, "step": 6199 }, { "epoch": 0.6700529557981195, "grad_norm": 0.44838985800743103, "learning_rate": 8e-05, "loss": 1.7523, "step": 6200 }, { "epoch": 0.6701610288555063, "grad_norm": 0.4209872782230377, "learning_rate": 8e-05, "loss": 1.7043, "step": 6201 }, { "epoch": 0.6702691019128931, "grad_norm": 0.42929089069366455, "learning_rate": 8e-05, "loss": 1.8322, "step": 6202 }, { "epoch": 0.6703771749702799, "grad_norm": 0.4010274112224579, "learning_rate": 8e-05, "loss": 1.7268, "step": 6203 }, { "epoch": 0.6704852480276667, "grad_norm": 0.45257872343063354, "learning_rate": 8e-05, "loss": 1.7566, "step": 6204 }, { "epoch": 0.6705933210850535, "grad_norm": 0.4022288918495178, "learning_rate": 8e-05, "loss": 1.6879, "step": 6205 }, { "epoch": 0.6707013941424403, "grad_norm": 0.4793477952480316, "learning_rate": 8e-05, "loss": 1.852, "step": 6206 }, { "epoch": 0.670809467199827, "grad_norm": 0.43653345108032227, "learning_rate": 8e-05, "loss": 1.7977, "step": 6207 }, { "epoch": 0.6709175402572138, "grad_norm": 0.42128685116767883, "learning_rate": 8e-05, "loss": 1.7287, "step": 6208 }, { "epoch": 0.6710256133146006, "grad_norm": 0.4590143859386444, "learning_rate": 8e-05, "loss": 1.8322, "step": 6209 }, { "epoch": 0.6711336863719874, "grad_norm": 0.46526455879211426, "learning_rate": 8e-05, "loss": 1.677, "step": 6210 }, { "epoch": 0.6712417594293743, "grad_norm": 0.4165186583995819, "learning_rate": 8e-05, "loss": 1.7277, "step": 6211 }, { "epoch": 0.6713498324867611, "grad_norm": 0.4142811894416809, "learning_rate": 8e-05, "loss": 1.6158, "step": 6212 }, { "epoch": 0.6714579055441479, "grad_norm": 0.3950520157814026, "learning_rate": 8e-05, "loss": 1.469, "step": 6213 }, { "epoch": 0.6715659786015347, "grad_norm": 0.4421716034412384, "learning_rate": 8e-05, "loss": 1.7762, "step": 6214 }, { "epoch": 0.6716740516589215, "grad_norm": 0.4067382514476776, "learning_rate": 8e-05, "loss": 1.645, "step": 6215 }, { "epoch": 0.6717821247163083, "grad_norm": 0.4422033727169037, "learning_rate": 8e-05, "loss": 1.6664, "step": 6216 }, { "epoch": 0.671890197773695, "grad_norm": 0.44875383377075195, "learning_rate": 8e-05, "loss": 1.7508, "step": 6217 }, { "epoch": 0.6719982708310818, "grad_norm": 0.4267762005329132, "learning_rate": 8e-05, "loss": 1.6955, "step": 6218 }, { "epoch": 0.6721063438884686, "grad_norm": 0.4368624687194824, "learning_rate": 8e-05, "loss": 1.7368, "step": 6219 }, { "epoch": 0.6722144169458554, "grad_norm": 0.4079696536064148, "learning_rate": 8e-05, "loss": 1.4867, "step": 6220 }, { "epoch": 0.6723224900032422, "grad_norm": 0.45279037952423096, "learning_rate": 8e-05, "loss": 1.6938, "step": 6221 }, { "epoch": 0.672430563060629, "grad_norm": 0.4431772530078888, "learning_rate": 8e-05, "loss": 1.755, "step": 6222 }, { "epoch": 0.6725386361180158, "grad_norm": 0.47114959359169006, "learning_rate": 8e-05, "loss": 1.6678, "step": 6223 }, { "epoch": 0.6726467091754026, "grad_norm": 0.5043656826019287, "learning_rate": 8e-05, "loss": 1.7492, "step": 6224 }, { "epoch": 0.6727547822327894, "grad_norm": 0.41463929414749146, "learning_rate": 8e-05, "loss": 1.5554, "step": 6225 }, { "epoch": 0.6728628552901762, "grad_norm": 0.44782665371894836, "learning_rate": 8e-05, "loss": 1.7914, "step": 6226 }, { "epoch": 0.672970928347563, "grad_norm": 0.4054979979991913, "learning_rate": 8e-05, "loss": 1.7258, "step": 6227 }, { "epoch": 0.6730790014049497, "grad_norm": 0.42032739520072937, "learning_rate": 8e-05, "loss": 1.6982, "step": 6228 }, { "epoch": 0.6731870744623365, "grad_norm": 0.4150976240634918, "learning_rate": 8e-05, "loss": 1.6263, "step": 6229 }, { "epoch": 0.6732951475197233, "grad_norm": 0.45023584365844727, "learning_rate": 8e-05, "loss": 1.8179, "step": 6230 }, { "epoch": 0.6734032205771101, "grad_norm": 0.4273819625377655, "learning_rate": 8e-05, "loss": 1.4635, "step": 6231 }, { "epoch": 0.6735112936344969, "grad_norm": 0.428329735994339, "learning_rate": 8e-05, "loss": 1.8206, "step": 6232 }, { "epoch": 0.6736193666918837, "grad_norm": 0.4503035247325897, "learning_rate": 8e-05, "loss": 1.7166, "step": 6233 }, { "epoch": 0.6737274397492705, "grad_norm": 0.41133883595466614, "learning_rate": 8e-05, "loss": 1.5766, "step": 6234 }, { "epoch": 0.6738355128066573, "grad_norm": 0.39508122205734253, "learning_rate": 8e-05, "loss": 1.649, "step": 6235 }, { "epoch": 0.6739435858640441, "grad_norm": 0.44302627444267273, "learning_rate": 8e-05, "loss": 1.762, "step": 6236 }, { "epoch": 0.6740516589214309, "grad_norm": 0.42008620500564575, "learning_rate": 8e-05, "loss": 1.6818, "step": 6237 }, { "epoch": 0.6741597319788177, "grad_norm": 0.39591455459594727, "learning_rate": 8e-05, "loss": 1.7235, "step": 6238 }, { "epoch": 0.6742678050362044, "grad_norm": 0.4631766378879547, "learning_rate": 8e-05, "loss": 1.6663, "step": 6239 }, { "epoch": 0.6743758780935912, "grad_norm": 0.44419610500335693, "learning_rate": 8e-05, "loss": 1.7731, "step": 6240 }, { "epoch": 0.674483951150978, "grad_norm": 0.568204402923584, "learning_rate": 8e-05, "loss": 1.8242, "step": 6241 }, { "epoch": 0.6745920242083648, "grad_norm": 0.43294504284858704, "learning_rate": 8e-05, "loss": 1.6585, "step": 6242 }, { "epoch": 0.6747000972657516, "grad_norm": 0.4084967076778412, "learning_rate": 8e-05, "loss": 1.7026, "step": 6243 }, { "epoch": 0.6748081703231384, "grad_norm": 0.4139155149459839, "learning_rate": 8e-05, "loss": 1.5357, "step": 6244 }, { "epoch": 0.6749162433805252, "grad_norm": 0.4834338128566742, "learning_rate": 8e-05, "loss": 1.7947, "step": 6245 }, { "epoch": 0.675024316437912, "grad_norm": 0.39736315608024597, "learning_rate": 8e-05, "loss": 1.5518, "step": 6246 }, { "epoch": 0.6751323894952989, "grad_norm": 0.44460031390190125, "learning_rate": 8e-05, "loss": 1.7609, "step": 6247 }, { "epoch": 0.6752404625526857, "grad_norm": 0.4076494872570038, "learning_rate": 8e-05, "loss": 1.7642, "step": 6248 }, { "epoch": 0.6753485356100724, "grad_norm": 0.44520390033721924, "learning_rate": 8e-05, "loss": 1.6886, "step": 6249 }, { "epoch": 0.6754566086674592, "grad_norm": 0.42820775508880615, "learning_rate": 8e-05, "loss": 1.6397, "step": 6250 }, { "epoch": 0.675564681724846, "grad_norm": 0.42849329113960266, "learning_rate": 8e-05, "loss": 1.732, "step": 6251 }, { "epoch": 0.6756727547822328, "grad_norm": 0.4788239300251007, "learning_rate": 8e-05, "loss": 1.9182, "step": 6252 }, { "epoch": 0.6757808278396196, "grad_norm": 0.42145049571990967, "learning_rate": 8e-05, "loss": 1.6454, "step": 6253 }, { "epoch": 0.6758889008970064, "grad_norm": 0.3814220726490021, "learning_rate": 8e-05, "loss": 1.5163, "step": 6254 }, { "epoch": 0.6759969739543932, "grad_norm": 0.43088364601135254, "learning_rate": 8e-05, "loss": 1.6782, "step": 6255 }, { "epoch": 0.67610504701178, "grad_norm": 0.45486611127853394, "learning_rate": 8e-05, "loss": 1.7404, "step": 6256 }, { "epoch": 0.6762131200691668, "grad_norm": 0.43307560682296753, "learning_rate": 8e-05, "loss": 1.691, "step": 6257 }, { "epoch": 0.6763211931265536, "grad_norm": 0.4152584671974182, "learning_rate": 8e-05, "loss": 1.6928, "step": 6258 }, { "epoch": 0.6764292661839404, "grad_norm": 0.505347490310669, "learning_rate": 8e-05, "loss": 1.5246, "step": 6259 }, { "epoch": 0.6765373392413271, "grad_norm": 0.4762478172779083, "learning_rate": 8e-05, "loss": 1.6753, "step": 6260 }, { "epoch": 0.6766454122987139, "grad_norm": 0.44121429324150085, "learning_rate": 8e-05, "loss": 1.8824, "step": 6261 }, { "epoch": 0.6767534853561007, "grad_norm": 0.4168994426727295, "learning_rate": 8e-05, "loss": 1.6746, "step": 6262 }, { "epoch": 0.6768615584134875, "grad_norm": 0.423836886882782, "learning_rate": 8e-05, "loss": 1.704, "step": 6263 }, { "epoch": 0.6769696314708743, "grad_norm": 0.4017491936683655, "learning_rate": 8e-05, "loss": 1.6447, "step": 6264 }, { "epoch": 0.6770777045282611, "grad_norm": 0.42914053797721863, "learning_rate": 8e-05, "loss": 1.708, "step": 6265 }, { "epoch": 0.6771857775856479, "grad_norm": 0.42630553245544434, "learning_rate": 8e-05, "loss": 1.8745, "step": 6266 }, { "epoch": 0.6772938506430347, "grad_norm": 0.4478112459182739, "learning_rate": 8e-05, "loss": 1.765, "step": 6267 }, { "epoch": 0.6774019237004215, "grad_norm": 0.41880956292152405, "learning_rate": 8e-05, "loss": 1.5336, "step": 6268 }, { "epoch": 0.6775099967578083, "grad_norm": 0.46359196305274963, "learning_rate": 8e-05, "loss": 1.7396, "step": 6269 }, { "epoch": 0.6776180698151951, "grad_norm": 0.4143178164958954, "learning_rate": 8e-05, "loss": 1.6607, "step": 6270 }, { "epoch": 0.6777261428725818, "grad_norm": 0.401677668094635, "learning_rate": 8e-05, "loss": 1.4645, "step": 6271 }, { "epoch": 0.6778342159299686, "grad_norm": 0.40817996859550476, "learning_rate": 8e-05, "loss": 1.5714, "step": 6272 }, { "epoch": 0.6779422889873554, "grad_norm": 0.4184369444847107, "learning_rate": 8e-05, "loss": 1.7353, "step": 6273 }, { "epoch": 0.6780503620447422, "grad_norm": 0.4592624604701996, "learning_rate": 8e-05, "loss": 1.8323, "step": 6274 }, { "epoch": 0.678158435102129, "grad_norm": 0.43698012828826904, "learning_rate": 8e-05, "loss": 1.6576, "step": 6275 }, { "epoch": 0.6782665081595158, "grad_norm": 0.4259214699268341, "learning_rate": 8e-05, "loss": 1.8375, "step": 6276 }, { "epoch": 0.6783745812169026, "grad_norm": 0.43012675642967224, "learning_rate": 8e-05, "loss": 1.886, "step": 6277 }, { "epoch": 0.6784826542742894, "grad_norm": 0.37964311242103577, "learning_rate": 8e-05, "loss": 1.5747, "step": 6278 }, { "epoch": 0.6785907273316762, "grad_norm": 0.41840481758117676, "learning_rate": 8e-05, "loss": 1.6732, "step": 6279 }, { "epoch": 0.678698800389063, "grad_norm": 0.4340895414352417, "learning_rate": 8e-05, "loss": 1.7654, "step": 6280 }, { "epoch": 0.6788068734464499, "grad_norm": 0.4496678113937378, "learning_rate": 8e-05, "loss": 1.7271, "step": 6281 }, { "epoch": 0.6789149465038365, "grad_norm": 0.41855761408805847, "learning_rate": 8e-05, "loss": 1.754, "step": 6282 }, { "epoch": 0.6790230195612234, "grad_norm": 0.41720765829086304, "learning_rate": 8e-05, "loss": 1.5937, "step": 6283 }, { "epoch": 0.6791310926186102, "grad_norm": 0.4091901481151581, "learning_rate": 8e-05, "loss": 1.6723, "step": 6284 }, { "epoch": 0.679239165675997, "grad_norm": 0.4083050787448883, "learning_rate": 8e-05, "loss": 1.6535, "step": 6285 }, { "epoch": 0.6793472387333838, "grad_norm": 0.42954546213150024, "learning_rate": 8e-05, "loss": 1.637, "step": 6286 }, { "epoch": 0.6794553117907706, "grad_norm": 0.4207683801651001, "learning_rate": 8e-05, "loss": 1.7429, "step": 6287 }, { "epoch": 0.6795633848481574, "grad_norm": 0.44910287857055664, "learning_rate": 8e-05, "loss": 1.6816, "step": 6288 }, { "epoch": 0.6796714579055442, "grad_norm": 0.46651551127433777, "learning_rate": 8e-05, "loss": 1.6612, "step": 6289 }, { "epoch": 0.679779530962931, "grad_norm": 0.40601515769958496, "learning_rate": 8e-05, "loss": 1.576, "step": 6290 }, { "epoch": 0.6798876040203178, "grad_norm": 0.41406336426734924, "learning_rate": 8e-05, "loss": 1.7922, "step": 6291 }, { "epoch": 0.6799956770777045, "grad_norm": 0.4611225128173828, "learning_rate": 8e-05, "loss": 1.7453, "step": 6292 }, { "epoch": 0.6801037501350913, "grad_norm": 0.3984382152557373, "learning_rate": 8e-05, "loss": 1.6288, "step": 6293 }, { "epoch": 0.6802118231924781, "grad_norm": 0.4365612864494324, "learning_rate": 8e-05, "loss": 1.7476, "step": 6294 }, { "epoch": 0.6803198962498649, "grad_norm": 0.44154518842697144, "learning_rate": 8e-05, "loss": 1.7617, "step": 6295 }, { "epoch": 0.6804279693072517, "grad_norm": 0.456993043422699, "learning_rate": 8e-05, "loss": 1.886, "step": 6296 }, { "epoch": 0.6805360423646385, "grad_norm": 0.4108124077320099, "learning_rate": 8e-05, "loss": 1.6446, "step": 6297 }, { "epoch": 0.6806441154220253, "grad_norm": 0.40727153420448303, "learning_rate": 8e-05, "loss": 1.6652, "step": 6298 }, { "epoch": 0.6807521884794121, "grad_norm": 0.44205716252326965, "learning_rate": 8e-05, "loss": 1.7396, "step": 6299 }, { "epoch": 0.6808602615367989, "grad_norm": 0.463917076587677, "learning_rate": 8e-05, "loss": 1.7514, "step": 6300 }, { "epoch": 0.6809683345941857, "grad_norm": 0.44297051429748535, "learning_rate": 8e-05, "loss": 1.7561, "step": 6301 }, { "epoch": 0.6810764076515725, "grad_norm": 0.485156387090683, "learning_rate": 8e-05, "loss": 1.7838, "step": 6302 }, { "epoch": 0.6811844807089592, "grad_norm": 0.4302632808685303, "learning_rate": 8e-05, "loss": 1.728, "step": 6303 }, { "epoch": 0.681292553766346, "grad_norm": 0.4471275806427002, "learning_rate": 8e-05, "loss": 1.761, "step": 6304 }, { "epoch": 0.6814006268237328, "grad_norm": 0.3929503560066223, "learning_rate": 8e-05, "loss": 1.6596, "step": 6305 }, { "epoch": 0.6815086998811196, "grad_norm": 0.4466809630393982, "learning_rate": 8e-05, "loss": 1.5088, "step": 6306 }, { "epoch": 0.6816167729385064, "grad_norm": 0.4596024751663208, "learning_rate": 8e-05, "loss": 1.7559, "step": 6307 }, { "epoch": 0.6817248459958932, "grad_norm": 0.40452784299850464, "learning_rate": 8e-05, "loss": 1.5824, "step": 6308 }, { "epoch": 0.68183291905328, "grad_norm": 0.42940354347229004, "learning_rate": 8e-05, "loss": 1.6775, "step": 6309 }, { "epoch": 0.6819409921106668, "grad_norm": 0.43611064553260803, "learning_rate": 8e-05, "loss": 1.6981, "step": 6310 }, { "epoch": 0.6820490651680536, "grad_norm": 0.4354825019836426, "learning_rate": 8e-05, "loss": 1.7884, "step": 6311 }, { "epoch": 0.6821571382254404, "grad_norm": 0.4024370312690735, "learning_rate": 8e-05, "loss": 1.6192, "step": 6312 }, { "epoch": 0.6822652112828272, "grad_norm": 0.4412800669670105, "learning_rate": 8e-05, "loss": 1.6359, "step": 6313 }, { "epoch": 0.6823732843402139, "grad_norm": 0.41143327951431274, "learning_rate": 8e-05, "loss": 1.759, "step": 6314 }, { "epoch": 0.6824813573976007, "grad_norm": 0.40496525168418884, "learning_rate": 8e-05, "loss": 1.6038, "step": 6315 }, { "epoch": 0.6825894304549875, "grad_norm": 0.4296342432498932, "learning_rate": 8e-05, "loss": 1.7049, "step": 6316 }, { "epoch": 0.6826975035123743, "grad_norm": 0.4329476058483124, "learning_rate": 8e-05, "loss": 1.7401, "step": 6317 }, { "epoch": 0.6828055765697612, "grad_norm": 0.4527119994163513, "learning_rate": 8e-05, "loss": 1.8207, "step": 6318 }, { "epoch": 0.682913649627148, "grad_norm": 0.4666050672531128, "learning_rate": 8e-05, "loss": 1.8786, "step": 6319 }, { "epoch": 0.6830217226845348, "grad_norm": 0.42124873399734497, "learning_rate": 8e-05, "loss": 1.6603, "step": 6320 }, { "epoch": 0.6831297957419216, "grad_norm": 0.40716809034347534, "learning_rate": 8e-05, "loss": 1.5647, "step": 6321 }, { "epoch": 0.6832378687993084, "grad_norm": 0.40731334686279297, "learning_rate": 8e-05, "loss": 1.5284, "step": 6322 }, { "epoch": 0.6833459418566952, "grad_norm": 0.5500004291534424, "learning_rate": 8e-05, "loss": 2.0882, "step": 6323 }, { "epoch": 0.6834540149140819, "grad_norm": 0.4404851794242859, "learning_rate": 8e-05, "loss": 1.6732, "step": 6324 }, { "epoch": 0.6835620879714687, "grad_norm": 0.488295316696167, "learning_rate": 8e-05, "loss": 1.6601, "step": 6325 }, { "epoch": 0.6836701610288555, "grad_norm": 0.4322751760482788, "learning_rate": 8e-05, "loss": 1.662, "step": 6326 }, { "epoch": 0.6837782340862423, "grad_norm": 0.4441797733306885, "learning_rate": 8e-05, "loss": 1.7152, "step": 6327 }, { "epoch": 0.6838863071436291, "grad_norm": 0.4210951328277588, "learning_rate": 8e-05, "loss": 1.632, "step": 6328 }, { "epoch": 0.6839943802010159, "grad_norm": 0.41117924451828003, "learning_rate": 8e-05, "loss": 1.5315, "step": 6329 }, { "epoch": 0.6841024532584027, "grad_norm": 0.4276168942451477, "learning_rate": 8e-05, "loss": 1.7359, "step": 6330 }, { "epoch": 0.6842105263157895, "grad_norm": 0.43006983399391174, "learning_rate": 8e-05, "loss": 1.6517, "step": 6331 }, { "epoch": 0.6843185993731763, "grad_norm": 0.4527553617954254, "learning_rate": 8e-05, "loss": 1.6756, "step": 6332 }, { "epoch": 0.6844266724305631, "grad_norm": 0.4648573100566864, "learning_rate": 8e-05, "loss": 1.9525, "step": 6333 }, { "epoch": 0.6845347454879499, "grad_norm": 0.41800010204315186, "learning_rate": 8e-05, "loss": 1.5644, "step": 6334 }, { "epoch": 0.6846428185453366, "grad_norm": 0.43238216638565063, "learning_rate": 8e-05, "loss": 1.505, "step": 6335 }, { "epoch": 0.6847508916027234, "grad_norm": 0.4429517984390259, "learning_rate": 8e-05, "loss": 1.7337, "step": 6336 }, { "epoch": 0.6848589646601102, "grad_norm": 0.4316408038139343, "learning_rate": 8e-05, "loss": 1.719, "step": 6337 }, { "epoch": 0.684967037717497, "grad_norm": 0.48034176230430603, "learning_rate": 8e-05, "loss": 1.7468, "step": 6338 }, { "epoch": 0.6850751107748838, "grad_norm": 0.40770453214645386, "learning_rate": 8e-05, "loss": 1.7516, "step": 6339 }, { "epoch": 0.6851831838322706, "grad_norm": 0.4707203805446625, "learning_rate": 8e-05, "loss": 1.7357, "step": 6340 }, { "epoch": 0.6852912568896574, "grad_norm": 0.43383851647377014, "learning_rate": 8e-05, "loss": 1.6907, "step": 6341 }, { "epoch": 0.6853993299470442, "grad_norm": 0.4945789873600006, "learning_rate": 8e-05, "loss": 1.9035, "step": 6342 }, { "epoch": 0.685507403004431, "grad_norm": 0.3943396508693695, "learning_rate": 8e-05, "loss": 1.6887, "step": 6343 }, { "epoch": 0.6856154760618178, "grad_norm": 0.39452865719795227, "learning_rate": 8e-05, "loss": 1.6212, "step": 6344 }, { "epoch": 0.6857235491192046, "grad_norm": 0.43430447578430176, "learning_rate": 8e-05, "loss": 1.7705, "step": 6345 }, { "epoch": 0.6858316221765913, "grad_norm": 0.454370379447937, "learning_rate": 8e-05, "loss": 1.7042, "step": 6346 }, { "epoch": 0.6859396952339781, "grad_norm": 0.4293418526649475, "learning_rate": 8e-05, "loss": 1.7898, "step": 6347 }, { "epoch": 0.6860477682913649, "grad_norm": 0.43156930804252625, "learning_rate": 8e-05, "loss": 1.649, "step": 6348 }, { "epoch": 0.6861558413487517, "grad_norm": 0.49769121408462524, "learning_rate": 8e-05, "loss": 1.824, "step": 6349 }, { "epoch": 0.6862639144061385, "grad_norm": 0.44062340259552, "learning_rate": 8e-05, "loss": 1.7322, "step": 6350 }, { "epoch": 0.6863719874635253, "grad_norm": 0.4359250068664551, "learning_rate": 8e-05, "loss": 1.7069, "step": 6351 }, { "epoch": 0.6864800605209121, "grad_norm": 0.45233458280563354, "learning_rate": 8e-05, "loss": 1.6584, "step": 6352 }, { "epoch": 0.686588133578299, "grad_norm": 0.42517584562301636, "learning_rate": 8e-05, "loss": 1.8011, "step": 6353 }, { "epoch": 0.6866962066356858, "grad_norm": 0.42120447754859924, "learning_rate": 8e-05, "loss": 1.641, "step": 6354 }, { "epoch": 0.6868042796930726, "grad_norm": 0.4475250244140625, "learning_rate": 8e-05, "loss": 1.6334, "step": 6355 }, { "epoch": 0.6869123527504594, "grad_norm": 0.43726688623428345, "learning_rate": 8e-05, "loss": 1.5324, "step": 6356 }, { "epoch": 0.6870204258078461, "grad_norm": 0.4196942150592804, "learning_rate": 8e-05, "loss": 1.6702, "step": 6357 }, { "epoch": 0.6871284988652329, "grad_norm": 0.43933403491973877, "learning_rate": 8e-05, "loss": 1.7731, "step": 6358 }, { "epoch": 0.6872365719226197, "grad_norm": 0.42578890919685364, "learning_rate": 8e-05, "loss": 1.7538, "step": 6359 }, { "epoch": 0.6873446449800065, "grad_norm": 0.3963402211666107, "learning_rate": 8e-05, "loss": 1.5783, "step": 6360 }, { "epoch": 0.6874527180373933, "grad_norm": 0.4237794876098633, "learning_rate": 8e-05, "loss": 1.686, "step": 6361 }, { "epoch": 0.6875607910947801, "grad_norm": 0.39096635580062866, "learning_rate": 8e-05, "loss": 1.7238, "step": 6362 }, { "epoch": 0.6876688641521669, "grad_norm": 0.41831475496292114, "learning_rate": 8e-05, "loss": 1.6486, "step": 6363 }, { "epoch": 0.6877769372095537, "grad_norm": 0.4197748303413391, "learning_rate": 8e-05, "loss": 1.8228, "step": 6364 }, { "epoch": 0.6878850102669405, "grad_norm": 0.4178147614002228, "learning_rate": 8e-05, "loss": 1.8154, "step": 6365 }, { "epoch": 0.6879930833243273, "grad_norm": 0.4066941440105438, "learning_rate": 8e-05, "loss": 1.8377, "step": 6366 }, { "epoch": 0.688101156381714, "grad_norm": 0.39582476019859314, "learning_rate": 8e-05, "loss": 1.6285, "step": 6367 }, { "epoch": 0.6882092294391008, "grad_norm": 0.41278761625289917, "learning_rate": 8e-05, "loss": 1.8244, "step": 6368 }, { "epoch": 0.6883173024964876, "grad_norm": 0.43042054772377014, "learning_rate": 8e-05, "loss": 1.6613, "step": 6369 }, { "epoch": 0.6884253755538744, "grad_norm": 0.3941178321838379, "learning_rate": 8e-05, "loss": 1.6703, "step": 6370 }, { "epoch": 0.6885334486112612, "grad_norm": 0.4373432695865631, "learning_rate": 8e-05, "loss": 1.7845, "step": 6371 }, { "epoch": 0.688641521668648, "grad_norm": 0.4604814648628235, "learning_rate": 8e-05, "loss": 1.7758, "step": 6372 }, { "epoch": 0.6887495947260348, "grad_norm": 0.4051070511341095, "learning_rate": 8e-05, "loss": 1.7316, "step": 6373 }, { "epoch": 0.6888576677834216, "grad_norm": 0.4129032492637634, "learning_rate": 8e-05, "loss": 1.7248, "step": 6374 }, { "epoch": 0.6889657408408084, "grad_norm": 0.41679078340530396, "learning_rate": 8e-05, "loss": 1.6396, "step": 6375 }, { "epoch": 0.6890738138981952, "grad_norm": 0.46112290024757385, "learning_rate": 8e-05, "loss": 1.9684, "step": 6376 }, { "epoch": 0.689181886955582, "grad_norm": 0.45056450366973877, "learning_rate": 8e-05, "loss": 1.5301, "step": 6377 }, { "epoch": 0.6892899600129687, "grad_norm": 0.4053666591644287, "learning_rate": 8e-05, "loss": 1.6628, "step": 6378 }, { "epoch": 0.6893980330703555, "grad_norm": 0.47929203510284424, "learning_rate": 8e-05, "loss": 1.9069, "step": 6379 }, { "epoch": 0.6895061061277423, "grad_norm": 0.4533064067363739, "learning_rate": 8e-05, "loss": 1.7707, "step": 6380 }, { "epoch": 0.6896141791851291, "grad_norm": 0.40223899483680725, "learning_rate": 8e-05, "loss": 1.6689, "step": 6381 }, { "epoch": 0.6897222522425159, "grad_norm": 0.47634491324424744, "learning_rate": 8e-05, "loss": 1.734, "step": 6382 }, { "epoch": 0.6898303252999027, "grad_norm": 0.4081016778945923, "learning_rate": 8e-05, "loss": 1.7356, "step": 6383 }, { "epoch": 0.6899383983572895, "grad_norm": 0.4351406991481781, "learning_rate": 8e-05, "loss": 1.7428, "step": 6384 }, { "epoch": 0.6900464714146763, "grad_norm": 0.4856329560279846, "learning_rate": 8e-05, "loss": 1.8174, "step": 6385 }, { "epoch": 0.6901545444720631, "grad_norm": 0.4375387728214264, "learning_rate": 8e-05, "loss": 1.5342, "step": 6386 }, { "epoch": 0.69026261752945, "grad_norm": 0.4159436821937561, "learning_rate": 8e-05, "loss": 1.6248, "step": 6387 }, { "epoch": 0.6903706905868368, "grad_norm": 0.3965224027633667, "learning_rate": 8e-05, "loss": 1.5549, "step": 6388 }, { "epoch": 0.6904787636442234, "grad_norm": 0.47170010209083557, "learning_rate": 8e-05, "loss": 1.7486, "step": 6389 }, { "epoch": 0.6905868367016103, "grad_norm": 0.4579146206378937, "learning_rate": 8e-05, "loss": 1.8, "step": 6390 }, { "epoch": 0.690694909758997, "grad_norm": 0.5616297721862793, "learning_rate": 8e-05, "loss": 1.6574, "step": 6391 }, { "epoch": 0.6908029828163839, "grad_norm": 0.4694068133831024, "learning_rate": 8e-05, "loss": 1.7301, "step": 6392 }, { "epoch": 0.6909110558737707, "grad_norm": 0.42115333676338196, "learning_rate": 8e-05, "loss": 1.6686, "step": 6393 }, { "epoch": 0.6910191289311575, "grad_norm": 0.41695478558540344, "learning_rate": 8e-05, "loss": 1.6643, "step": 6394 }, { "epoch": 0.6911272019885443, "grad_norm": 0.4747338891029358, "learning_rate": 8e-05, "loss": 1.7438, "step": 6395 }, { "epoch": 0.6912352750459311, "grad_norm": 0.42129960656166077, "learning_rate": 8e-05, "loss": 1.6981, "step": 6396 }, { "epoch": 0.6913433481033179, "grad_norm": 0.4250677824020386, "learning_rate": 8e-05, "loss": 1.6461, "step": 6397 }, { "epoch": 0.6914514211607047, "grad_norm": 0.42111313343048096, "learning_rate": 8e-05, "loss": 1.6988, "step": 6398 }, { "epoch": 0.6915594942180914, "grad_norm": 0.49102962017059326, "learning_rate": 8e-05, "loss": 1.7987, "step": 6399 }, { "epoch": 0.6916675672754782, "grad_norm": 0.4287731647491455, "learning_rate": 8e-05, "loss": 1.8393, "step": 6400 }, { "epoch": 0.691775640332865, "grad_norm": 0.4392477571964264, "learning_rate": 8e-05, "loss": 1.7094, "step": 6401 }, { "epoch": 0.6918837133902518, "grad_norm": 0.39526450634002686, "learning_rate": 8e-05, "loss": 1.5063, "step": 6402 }, { "epoch": 0.6919917864476386, "grad_norm": 0.41698625683784485, "learning_rate": 8e-05, "loss": 1.7119, "step": 6403 }, { "epoch": 0.6920998595050254, "grad_norm": 0.4175799787044525, "learning_rate": 8e-05, "loss": 1.7707, "step": 6404 }, { "epoch": 0.6922079325624122, "grad_norm": 0.39970511198043823, "learning_rate": 8e-05, "loss": 1.4768, "step": 6405 }, { "epoch": 0.692316005619799, "grad_norm": 0.4526476562023163, "learning_rate": 8e-05, "loss": 1.5165, "step": 6406 }, { "epoch": 0.6924240786771858, "grad_norm": 0.46272042393684387, "learning_rate": 8e-05, "loss": 1.6512, "step": 6407 }, { "epoch": 0.6925321517345726, "grad_norm": 0.4754316210746765, "learning_rate": 8e-05, "loss": 1.8912, "step": 6408 }, { "epoch": 0.6926402247919594, "grad_norm": 0.5288708806037903, "learning_rate": 8e-05, "loss": 1.7232, "step": 6409 }, { "epoch": 0.6927482978493461, "grad_norm": 0.4434646666049957, "learning_rate": 8e-05, "loss": 1.8572, "step": 6410 }, { "epoch": 0.6928563709067329, "grad_norm": 0.46261531114578247, "learning_rate": 8e-05, "loss": 1.8014, "step": 6411 }, { "epoch": 0.6929644439641197, "grad_norm": 0.4744204878807068, "learning_rate": 8e-05, "loss": 1.8523, "step": 6412 }, { "epoch": 0.6930725170215065, "grad_norm": 0.46904417872428894, "learning_rate": 8e-05, "loss": 1.8557, "step": 6413 }, { "epoch": 0.6931805900788933, "grad_norm": 0.4081658720970154, "learning_rate": 8e-05, "loss": 1.6901, "step": 6414 }, { "epoch": 0.6932886631362801, "grad_norm": 0.4504369795322418, "learning_rate": 8e-05, "loss": 1.7804, "step": 6415 }, { "epoch": 0.6933967361936669, "grad_norm": 0.4169376790523529, "learning_rate": 8e-05, "loss": 1.6275, "step": 6416 }, { "epoch": 0.6935048092510537, "grad_norm": 0.451588898897171, "learning_rate": 8e-05, "loss": 1.766, "step": 6417 }, { "epoch": 0.6936128823084405, "grad_norm": 0.43449947237968445, "learning_rate": 8e-05, "loss": 1.6767, "step": 6418 }, { "epoch": 0.6937209553658273, "grad_norm": 0.41969114542007446, "learning_rate": 8e-05, "loss": 1.6624, "step": 6419 }, { "epoch": 0.6938290284232141, "grad_norm": 0.41564515233039856, "learning_rate": 8e-05, "loss": 1.6581, "step": 6420 }, { "epoch": 0.6939371014806008, "grad_norm": 0.390701025724411, "learning_rate": 8e-05, "loss": 1.5832, "step": 6421 }, { "epoch": 0.6940451745379876, "grad_norm": 0.4255692660808563, "learning_rate": 8e-05, "loss": 1.5868, "step": 6422 }, { "epoch": 0.6941532475953744, "grad_norm": 0.5021861791610718, "learning_rate": 8e-05, "loss": 1.7493, "step": 6423 }, { "epoch": 0.6942613206527612, "grad_norm": 0.43268901109695435, "learning_rate": 8e-05, "loss": 1.7409, "step": 6424 }, { "epoch": 0.694369393710148, "grad_norm": 0.4655124247074127, "learning_rate": 8e-05, "loss": 1.8337, "step": 6425 }, { "epoch": 0.6944774667675349, "grad_norm": 0.3918802738189697, "learning_rate": 8e-05, "loss": 1.5168, "step": 6426 }, { "epoch": 0.6945855398249217, "grad_norm": 0.3941533863544464, "learning_rate": 8e-05, "loss": 1.6012, "step": 6427 }, { "epoch": 0.6946936128823085, "grad_norm": 0.44900640845298767, "learning_rate": 8e-05, "loss": 1.5772, "step": 6428 }, { "epoch": 0.6948016859396953, "grad_norm": 0.4350531995296478, "learning_rate": 8e-05, "loss": 1.5988, "step": 6429 }, { "epoch": 0.6949097589970821, "grad_norm": 0.40637409687042236, "learning_rate": 8e-05, "loss": 1.6829, "step": 6430 }, { "epoch": 0.6950178320544689, "grad_norm": 0.4647524654865265, "learning_rate": 8e-05, "loss": 1.3453, "step": 6431 }, { "epoch": 0.6951259051118556, "grad_norm": 0.45278820395469666, "learning_rate": 8e-05, "loss": 1.6771, "step": 6432 }, { "epoch": 0.6952339781692424, "grad_norm": 0.4479602575302124, "learning_rate": 8e-05, "loss": 1.806, "step": 6433 }, { "epoch": 0.6953420512266292, "grad_norm": 0.41167354583740234, "learning_rate": 8e-05, "loss": 1.6752, "step": 6434 }, { "epoch": 0.695450124284016, "grad_norm": 0.4564761519432068, "learning_rate": 8e-05, "loss": 1.6769, "step": 6435 }, { "epoch": 0.6955581973414028, "grad_norm": 0.4306217133998871, "learning_rate": 8e-05, "loss": 1.6156, "step": 6436 }, { "epoch": 0.6956662703987896, "grad_norm": 0.4619109034538269, "learning_rate": 8e-05, "loss": 1.8106, "step": 6437 }, { "epoch": 0.6957743434561764, "grad_norm": 0.40540438890457153, "learning_rate": 8e-05, "loss": 1.5321, "step": 6438 }, { "epoch": 0.6958824165135632, "grad_norm": 0.45563438534736633, "learning_rate": 8e-05, "loss": 1.7536, "step": 6439 }, { "epoch": 0.69599048957095, "grad_norm": 0.47033578157424927, "learning_rate": 8e-05, "loss": 1.5762, "step": 6440 }, { "epoch": 0.6960985626283368, "grad_norm": 0.4173518717288971, "learning_rate": 8e-05, "loss": 1.7024, "step": 6441 }, { "epoch": 0.6962066356857235, "grad_norm": 0.45252615213394165, "learning_rate": 8e-05, "loss": 1.8543, "step": 6442 }, { "epoch": 0.6963147087431103, "grad_norm": 0.4434443712234497, "learning_rate": 8e-05, "loss": 1.6226, "step": 6443 }, { "epoch": 0.6964227818004971, "grad_norm": 0.48867034912109375, "learning_rate": 8e-05, "loss": 1.7645, "step": 6444 }, { "epoch": 0.6965308548578839, "grad_norm": 0.442917138338089, "learning_rate": 8e-05, "loss": 1.7054, "step": 6445 }, { "epoch": 0.6966389279152707, "grad_norm": 0.4726044833660126, "learning_rate": 8e-05, "loss": 1.8281, "step": 6446 }, { "epoch": 0.6967470009726575, "grad_norm": 0.4536777138710022, "learning_rate": 8e-05, "loss": 1.6943, "step": 6447 }, { "epoch": 0.6968550740300443, "grad_norm": 0.4115740954875946, "learning_rate": 8e-05, "loss": 1.827, "step": 6448 }, { "epoch": 0.6969631470874311, "grad_norm": 0.4536694288253784, "learning_rate": 8e-05, "loss": 1.7819, "step": 6449 }, { "epoch": 0.6970712201448179, "grad_norm": 0.4217135012149811, "learning_rate": 8e-05, "loss": 1.6836, "step": 6450 }, { "epoch": 0.6971792932022047, "grad_norm": 0.4677889049053192, "learning_rate": 8e-05, "loss": 1.8653, "step": 6451 }, { "epoch": 0.6972873662595915, "grad_norm": 0.4256700277328491, "learning_rate": 8e-05, "loss": 1.7124, "step": 6452 }, { "epoch": 0.6973954393169782, "grad_norm": 0.5305623412132263, "learning_rate": 8e-05, "loss": 1.6674, "step": 6453 }, { "epoch": 0.697503512374365, "grad_norm": 0.4220481812953949, "learning_rate": 8e-05, "loss": 1.7062, "step": 6454 }, { "epoch": 0.6976115854317518, "grad_norm": 0.39415690302848816, "learning_rate": 8e-05, "loss": 1.5237, "step": 6455 }, { "epoch": 0.6977196584891386, "grad_norm": 0.442704975605011, "learning_rate": 8e-05, "loss": 1.8463, "step": 6456 }, { "epoch": 0.6978277315465254, "grad_norm": 0.382398396730423, "learning_rate": 8e-05, "loss": 1.421, "step": 6457 }, { "epoch": 0.6979358046039122, "grad_norm": 0.391323983669281, "learning_rate": 8e-05, "loss": 1.5574, "step": 6458 }, { "epoch": 0.698043877661299, "grad_norm": 0.47067055106163025, "learning_rate": 8e-05, "loss": 1.8034, "step": 6459 }, { "epoch": 0.6981519507186859, "grad_norm": 0.41287410259246826, "learning_rate": 8e-05, "loss": 1.8042, "step": 6460 }, { "epoch": 0.6982600237760727, "grad_norm": 0.40971919894218445, "learning_rate": 8e-05, "loss": 1.6702, "step": 6461 }, { "epoch": 0.6983680968334595, "grad_norm": 0.41088780760765076, "learning_rate": 8e-05, "loss": 1.6555, "step": 6462 }, { "epoch": 0.6984761698908463, "grad_norm": 0.40319857001304626, "learning_rate": 8e-05, "loss": 1.6089, "step": 6463 }, { "epoch": 0.698584242948233, "grad_norm": 0.4425109922885895, "learning_rate": 8e-05, "loss": 1.7525, "step": 6464 }, { "epoch": 0.6986923160056198, "grad_norm": 0.40813928842544556, "learning_rate": 8e-05, "loss": 1.7163, "step": 6465 }, { "epoch": 0.6988003890630066, "grad_norm": 0.41607627272605896, "learning_rate": 8e-05, "loss": 1.6928, "step": 6466 }, { "epoch": 0.6989084621203934, "grad_norm": 0.4172656536102295, "learning_rate": 8e-05, "loss": 1.6778, "step": 6467 }, { "epoch": 0.6990165351777802, "grad_norm": 0.41422805190086365, "learning_rate": 8e-05, "loss": 1.5875, "step": 6468 }, { "epoch": 0.699124608235167, "grad_norm": 0.45193955302238464, "learning_rate": 8e-05, "loss": 1.6783, "step": 6469 }, { "epoch": 0.6992326812925538, "grad_norm": 0.4057518541812897, "learning_rate": 8e-05, "loss": 1.7117, "step": 6470 }, { "epoch": 0.6993407543499406, "grad_norm": 0.4412029981613159, "learning_rate": 8e-05, "loss": 1.8378, "step": 6471 }, { "epoch": 0.6994488274073274, "grad_norm": 0.4311342239379883, "learning_rate": 8e-05, "loss": 1.6658, "step": 6472 }, { "epoch": 0.6995569004647142, "grad_norm": 0.4088904857635498, "learning_rate": 8e-05, "loss": 1.6788, "step": 6473 }, { "epoch": 0.6996649735221009, "grad_norm": 0.4973047971725464, "learning_rate": 8e-05, "loss": 1.8235, "step": 6474 }, { "epoch": 0.6997730465794877, "grad_norm": 0.45359349250793457, "learning_rate": 8e-05, "loss": 1.6865, "step": 6475 }, { "epoch": 0.6998811196368745, "grad_norm": 0.4033246338367462, "learning_rate": 8e-05, "loss": 1.5425, "step": 6476 }, { "epoch": 0.6999891926942613, "grad_norm": 0.40630337595939636, "learning_rate": 8e-05, "loss": 1.642, "step": 6477 }, { "epoch": 0.7000972657516481, "grad_norm": 0.4371485412120819, "learning_rate": 8e-05, "loss": 1.863, "step": 6478 }, { "epoch": 0.7002053388090349, "grad_norm": 0.4048522710800171, "learning_rate": 8e-05, "loss": 1.5431, "step": 6479 }, { "epoch": 0.7003134118664217, "grad_norm": 0.39445072412490845, "learning_rate": 8e-05, "loss": 1.6093, "step": 6480 }, { "epoch": 0.7004214849238085, "grad_norm": 0.4971485733985901, "learning_rate": 8e-05, "loss": 1.8939, "step": 6481 }, { "epoch": 0.7005295579811953, "grad_norm": 0.4782235026359558, "learning_rate": 8e-05, "loss": 1.9632, "step": 6482 }, { "epoch": 0.7006376310385821, "grad_norm": 0.4286333918571472, "learning_rate": 8e-05, "loss": 1.7718, "step": 6483 }, { "epoch": 0.7007457040959689, "grad_norm": 0.4708327054977417, "learning_rate": 8e-05, "loss": 1.812, "step": 6484 }, { "epoch": 0.7008537771533556, "grad_norm": 0.40467116236686707, "learning_rate": 8e-05, "loss": 1.7223, "step": 6485 }, { "epoch": 0.7009618502107424, "grad_norm": 0.4190639853477478, "learning_rate": 8e-05, "loss": 1.6957, "step": 6486 }, { "epoch": 0.7010699232681292, "grad_norm": 0.39088359475135803, "learning_rate": 8e-05, "loss": 1.5965, "step": 6487 }, { "epoch": 0.701177996325516, "grad_norm": 0.4225016236305237, "learning_rate": 8e-05, "loss": 1.6755, "step": 6488 }, { "epoch": 0.7012860693829028, "grad_norm": 0.42351222038269043, "learning_rate": 8e-05, "loss": 1.4913, "step": 6489 }, { "epoch": 0.7013941424402896, "grad_norm": 0.40386590361595154, "learning_rate": 8e-05, "loss": 1.6792, "step": 6490 }, { "epoch": 0.7015022154976764, "grad_norm": 0.4635016918182373, "learning_rate": 8e-05, "loss": 1.7924, "step": 6491 }, { "epoch": 0.7016102885550632, "grad_norm": 0.39641475677490234, "learning_rate": 8e-05, "loss": 1.5928, "step": 6492 }, { "epoch": 0.70171836161245, "grad_norm": 0.44617363810539246, "learning_rate": 8e-05, "loss": 1.8562, "step": 6493 }, { "epoch": 0.7018264346698369, "grad_norm": 0.4382881224155426, "learning_rate": 8e-05, "loss": 1.8294, "step": 6494 }, { "epoch": 0.7019345077272237, "grad_norm": 0.4407083988189697, "learning_rate": 8e-05, "loss": 1.75, "step": 6495 }, { "epoch": 0.7020425807846103, "grad_norm": 0.46347278356552124, "learning_rate": 8e-05, "loss": 1.8654, "step": 6496 }, { "epoch": 0.7021506538419972, "grad_norm": 0.462984174489975, "learning_rate": 8e-05, "loss": 1.6982, "step": 6497 }, { "epoch": 0.702258726899384, "grad_norm": 0.4184113144874573, "learning_rate": 8e-05, "loss": 1.8224, "step": 6498 }, { "epoch": 0.7023667999567708, "grad_norm": 0.47514721751213074, "learning_rate": 8e-05, "loss": 1.8251, "step": 6499 }, { "epoch": 0.7024748730141576, "grad_norm": 0.40523871779441833, "learning_rate": 8e-05, "loss": 1.6495, "step": 6500 }, { "epoch": 0.7025829460715444, "grad_norm": 0.43650874495506287, "learning_rate": 8e-05, "loss": 1.5732, "step": 6501 }, { "epoch": 0.7026910191289312, "grad_norm": 0.47068262100219727, "learning_rate": 8e-05, "loss": 1.784, "step": 6502 }, { "epoch": 0.702799092186318, "grad_norm": 0.40445953607559204, "learning_rate": 8e-05, "loss": 1.6147, "step": 6503 }, { "epoch": 0.7029071652437048, "grad_norm": 0.41558483242988586, "learning_rate": 8e-05, "loss": 1.7167, "step": 6504 }, { "epoch": 0.7030152383010916, "grad_norm": 0.43552669882774353, "learning_rate": 8e-05, "loss": 1.7005, "step": 6505 }, { "epoch": 0.7031233113584784, "grad_norm": 0.4935568571090698, "learning_rate": 8e-05, "loss": 1.7335, "step": 6506 }, { "epoch": 0.7032313844158651, "grad_norm": 0.4105222821235657, "learning_rate": 8e-05, "loss": 1.6714, "step": 6507 }, { "epoch": 0.7033394574732519, "grad_norm": 0.4174345135688782, "learning_rate": 8e-05, "loss": 1.7493, "step": 6508 }, { "epoch": 0.7034475305306387, "grad_norm": 0.40639981627464294, "learning_rate": 8e-05, "loss": 1.7436, "step": 6509 }, { "epoch": 0.7035556035880255, "grad_norm": 0.4689512848854065, "learning_rate": 8e-05, "loss": 1.5869, "step": 6510 }, { "epoch": 0.7036636766454123, "grad_norm": 0.4034641981124878, "learning_rate": 8e-05, "loss": 1.4024, "step": 6511 }, { "epoch": 0.7037717497027991, "grad_norm": 0.476927250623703, "learning_rate": 8e-05, "loss": 1.7143, "step": 6512 }, { "epoch": 0.7038798227601859, "grad_norm": 0.40448057651519775, "learning_rate": 8e-05, "loss": 1.6821, "step": 6513 }, { "epoch": 0.7039878958175727, "grad_norm": 0.4777764081954956, "learning_rate": 8e-05, "loss": 1.6868, "step": 6514 }, { "epoch": 0.7040959688749595, "grad_norm": 0.4432443380355835, "learning_rate": 8e-05, "loss": 1.8457, "step": 6515 }, { "epoch": 0.7042040419323463, "grad_norm": 0.41441816091537476, "learning_rate": 8e-05, "loss": 1.7485, "step": 6516 }, { "epoch": 0.704312114989733, "grad_norm": 0.46053338050842285, "learning_rate": 8e-05, "loss": 1.7922, "step": 6517 }, { "epoch": 0.7044201880471198, "grad_norm": 0.4405980408191681, "learning_rate": 8e-05, "loss": 1.604, "step": 6518 }, { "epoch": 0.7045282611045066, "grad_norm": 0.4100898504257202, "learning_rate": 8e-05, "loss": 1.5932, "step": 6519 }, { "epoch": 0.7046363341618934, "grad_norm": 0.46582862734794617, "learning_rate": 8e-05, "loss": 1.7576, "step": 6520 }, { "epoch": 0.7047444072192802, "grad_norm": 0.414947509765625, "learning_rate": 8e-05, "loss": 1.6011, "step": 6521 }, { "epoch": 0.704852480276667, "grad_norm": 0.4548057019710541, "learning_rate": 8e-05, "loss": 1.8765, "step": 6522 }, { "epoch": 0.7049605533340538, "grad_norm": 0.42565372586250305, "learning_rate": 8e-05, "loss": 1.6844, "step": 6523 }, { "epoch": 0.7050686263914406, "grad_norm": 0.41039445996284485, "learning_rate": 8e-05, "loss": 1.6333, "step": 6524 }, { "epoch": 0.7051766994488274, "grad_norm": 0.42569488286972046, "learning_rate": 8e-05, "loss": 1.7043, "step": 6525 }, { "epoch": 0.7052847725062142, "grad_norm": 0.45719969272613525, "learning_rate": 8e-05, "loss": 1.748, "step": 6526 }, { "epoch": 0.705392845563601, "grad_norm": 0.4595228433609009, "learning_rate": 8e-05, "loss": 1.6829, "step": 6527 }, { "epoch": 0.7055009186209877, "grad_norm": 0.41614681482315063, "learning_rate": 8e-05, "loss": 1.5444, "step": 6528 }, { "epoch": 0.7056089916783745, "grad_norm": 0.42489874362945557, "learning_rate": 8e-05, "loss": 1.6689, "step": 6529 }, { "epoch": 0.7057170647357613, "grad_norm": 0.43700963258743286, "learning_rate": 8e-05, "loss": 1.5206, "step": 6530 }, { "epoch": 0.7058251377931481, "grad_norm": 0.46417900919914246, "learning_rate": 8e-05, "loss": 1.7821, "step": 6531 }, { "epoch": 0.705933210850535, "grad_norm": 0.40817323327064514, "learning_rate": 8e-05, "loss": 1.7626, "step": 6532 }, { "epoch": 0.7060412839079218, "grad_norm": 0.4195951521396637, "learning_rate": 8e-05, "loss": 1.7832, "step": 6533 }, { "epoch": 0.7061493569653086, "grad_norm": 0.39281895756721497, "learning_rate": 8e-05, "loss": 1.5415, "step": 6534 }, { "epoch": 0.7062574300226954, "grad_norm": 0.44885316491127014, "learning_rate": 8e-05, "loss": 1.7526, "step": 6535 }, { "epoch": 0.7063655030800822, "grad_norm": 0.4100796580314636, "learning_rate": 8e-05, "loss": 1.7735, "step": 6536 }, { "epoch": 0.706473576137469, "grad_norm": 0.4301604628562927, "learning_rate": 8e-05, "loss": 1.7945, "step": 6537 }, { "epoch": 0.7065816491948558, "grad_norm": 0.43208739161491394, "learning_rate": 8e-05, "loss": 1.5593, "step": 6538 }, { "epoch": 0.7066897222522425, "grad_norm": 0.40398970246315, "learning_rate": 8e-05, "loss": 1.6265, "step": 6539 }, { "epoch": 0.7067977953096293, "grad_norm": 0.4335843622684479, "learning_rate": 8e-05, "loss": 1.8148, "step": 6540 }, { "epoch": 0.7069058683670161, "grad_norm": 0.4159396290779114, "learning_rate": 8e-05, "loss": 1.6424, "step": 6541 }, { "epoch": 0.7070139414244029, "grad_norm": 0.4156258702278137, "learning_rate": 8e-05, "loss": 1.7668, "step": 6542 }, { "epoch": 0.7071220144817897, "grad_norm": 0.4263935387134552, "learning_rate": 8e-05, "loss": 1.7156, "step": 6543 }, { "epoch": 0.7072300875391765, "grad_norm": 0.46826398372650146, "learning_rate": 8e-05, "loss": 1.763, "step": 6544 }, { "epoch": 0.7073381605965633, "grad_norm": 0.40507641434669495, "learning_rate": 8e-05, "loss": 1.6109, "step": 6545 }, { "epoch": 0.7074462336539501, "grad_norm": 0.40346115827560425, "learning_rate": 8e-05, "loss": 1.5457, "step": 6546 }, { "epoch": 0.7075543067113369, "grad_norm": 0.43222230672836304, "learning_rate": 8e-05, "loss": 1.6756, "step": 6547 }, { "epoch": 0.7076623797687237, "grad_norm": 0.4581906795501709, "learning_rate": 8e-05, "loss": 1.7266, "step": 6548 }, { "epoch": 0.7077704528261104, "grad_norm": 0.505719006061554, "learning_rate": 8e-05, "loss": 1.909, "step": 6549 }, { "epoch": 0.7078785258834972, "grad_norm": 0.3996567726135254, "learning_rate": 8e-05, "loss": 1.7235, "step": 6550 }, { "epoch": 0.707986598940884, "grad_norm": 0.4255920350551605, "learning_rate": 8e-05, "loss": 1.7934, "step": 6551 }, { "epoch": 0.7080946719982708, "grad_norm": 0.4288960099220276, "learning_rate": 8e-05, "loss": 1.5995, "step": 6552 }, { "epoch": 0.7082027450556576, "grad_norm": 0.41746410727500916, "learning_rate": 8e-05, "loss": 1.6938, "step": 6553 }, { "epoch": 0.7083108181130444, "grad_norm": 0.40016570687294006, "learning_rate": 8e-05, "loss": 1.4512, "step": 6554 }, { "epoch": 0.7084188911704312, "grad_norm": 0.39908507466316223, "learning_rate": 8e-05, "loss": 1.8006, "step": 6555 }, { "epoch": 0.708526964227818, "grad_norm": 0.3981580436229706, "learning_rate": 8e-05, "loss": 1.5936, "step": 6556 }, { "epoch": 0.7086350372852048, "grad_norm": 0.4299772381782532, "learning_rate": 8e-05, "loss": 1.66, "step": 6557 }, { "epoch": 0.7087431103425916, "grad_norm": 0.46663716435432434, "learning_rate": 8e-05, "loss": 1.8448, "step": 6558 }, { "epoch": 0.7088511833999784, "grad_norm": 0.4137229919433594, "learning_rate": 8e-05, "loss": 1.6424, "step": 6559 }, { "epoch": 0.7089592564573651, "grad_norm": 0.41194120049476624, "learning_rate": 8e-05, "loss": 1.56, "step": 6560 }, { "epoch": 0.7090673295147519, "grad_norm": 0.45921534299850464, "learning_rate": 8e-05, "loss": 1.4317, "step": 6561 }, { "epoch": 0.7091754025721387, "grad_norm": 0.4668531119823456, "learning_rate": 8e-05, "loss": 1.6044, "step": 6562 }, { "epoch": 0.7092834756295255, "grad_norm": 0.4660416841506958, "learning_rate": 8e-05, "loss": 1.7615, "step": 6563 }, { "epoch": 0.7093915486869123, "grad_norm": 0.4485475718975067, "learning_rate": 8e-05, "loss": 1.6861, "step": 6564 }, { "epoch": 0.7094996217442991, "grad_norm": 0.4287455677986145, "learning_rate": 8e-05, "loss": 1.6076, "step": 6565 }, { "epoch": 0.709607694801686, "grad_norm": 0.41513583064079285, "learning_rate": 8e-05, "loss": 1.6224, "step": 6566 }, { "epoch": 0.7097157678590728, "grad_norm": 0.3865039050579071, "learning_rate": 8e-05, "loss": 1.6692, "step": 6567 }, { "epoch": 0.7098238409164596, "grad_norm": 0.4492741823196411, "learning_rate": 8e-05, "loss": 1.7062, "step": 6568 }, { "epoch": 0.7099319139738464, "grad_norm": 0.41714730858802795, "learning_rate": 8e-05, "loss": 1.7196, "step": 6569 }, { "epoch": 0.7100399870312332, "grad_norm": 0.43117478489875793, "learning_rate": 8e-05, "loss": 1.6839, "step": 6570 }, { "epoch": 0.7101480600886199, "grad_norm": 0.41240814328193665, "learning_rate": 8e-05, "loss": 1.7461, "step": 6571 }, { "epoch": 0.7102561331460067, "grad_norm": 0.44664350152015686, "learning_rate": 8e-05, "loss": 1.7228, "step": 6572 }, { "epoch": 0.7103642062033935, "grad_norm": 0.4356743395328522, "learning_rate": 8e-05, "loss": 1.7391, "step": 6573 }, { "epoch": 0.7104722792607803, "grad_norm": 0.4056012034416199, "learning_rate": 8e-05, "loss": 1.692, "step": 6574 }, { "epoch": 0.7105803523181671, "grad_norm": 0.4359751045703888, "learning_rate": 8e-05, "loss": 1.5659, "step": 6575 }, { "epoch": 0.7106884253755539, "grad_norm": 0.4049901068210602, "learning_rate": 8e-05, "loss": 1.7083, "step": 6576 }, { "epoch": 0.7107964984329407, "grad_norm": 0.457628458738327, "learning_rate": 8e-05, "loss": 1.7872, "step": 6577 }, { "epoch": 0.7109045714903275, "grad_norm": 0.43600258231163025, "learning_rate": 8e-05, "loss": 1.7804, "step": 6578 }, { "epoch": 0.7110126445477143, "grad_norm": 0.44636744260787964, "learning_rate": 8e-05, "loss": 1.8464, "step": 6579 }, { "epoch": 0.7111207176051011, "grad_norm": 0.4442000389099121, "learning_rate": 8e-05, "loss": 1.6537, "step": 6580 }, { "epoch": 0.7112287906624878, "grad_norm": 0.4303392469882965, "learning_rate": 8e-05, "loss": 1.6634, "step": 6581 }, { "epoch": 0.7113368637198746, "grad_norm": 0.41447535157203674, "learning_rate": 8e-05, "loss": 1.4582, "step": 6582 }, { "epoch": 0.7114449367772614, "grad_norm": 0.3947119116783142, "learning_rate": 8e-05, "loss": 1.5328, "step": 6583 }, { "epoch": 0.7115530098346482, "grad_norm": 0.3833637237548828, "learning_rate": 8e-05, "loss": 1.5809, "step": 6584 }, { "epoch": 0.711661082892035, "grad_norm": 0.4587970972061157, "learning_rate": 8e-05, "loss": 1.7545, "step": 6585 }, { "epoch": 0.7117691559494218, "grad_norm": 0.4138355255126953, "learning_rate": 8e-05, "loss": 1.5807, "step": 6586 }, { "epoch": 0.7118772290068086, "grad_norm": 0.40345677733421326, "learning_rate": 8e-05, "loss": 1.6584, "step": 6587 }, { "epoch": 0.7119853020641954, "grad_norm": 0.41415050625801086, "learning_rate": 8e-05, "loss": 1.6444, "step": 6588 }, { "epoch": 0.7120933751215822, "grad_norm": 0.46794530749320984, "learning_rate": 8e-05, "loss": 1.907, "step": 6589 }, { "epoch": 0.712201448178969, "grad_norm": 0.4273184835910797, "learning_rate": 8e-05, "loss": 1.6515, "step": 6590 }, { "epoch": 0.7123095212363558, "grad_norm": 0.4163542687892914, "learning_rate": 8e-05, "loss": 1.7034, "step": 6591 }, { "epoch": 0.7124175942937425, "grad_norm": 0.46267786622047424, "learning_rate": 8e-05, "loss": 1.7764, "step": 6592 }, { "epoch": 0.7125256673511293, "grad_norm": 0.39994704723358154, "learning_rate": 8e-05, "loss": 1.5977, "step": 6593 }, { "epoch": 0.7126337404085161, "grad_norm": 0.42116111516952515, "learning_rate": 8e-05, "loss": 1.7034, "step": 6594 }, { "epoch": 0.7127418134659029, "grad_norm": 0.41757577657699585, "learning_rate": 8e-05, "loss": 1.6486, "step": 6595 }, { "epoch": 0.7128498865232897, "grad_norm": 0.5341874361038208, "learning_rate": 8e-05, "loss": 1.8287, "step": 6596 }, { "epoch": 0.7129579595806765, "grad_norm": 0.41664406657218933, "learning_rate": 8e-05, "loss": 1.7649, "step": 6597 }, { "epoch": 0.7130660326380633, "grad_norm": 0.4036864936351776, "learning_rate": 8e-05, "loss": 1.6935, "step": 6598 }, { "epoch": 0.7131741056954501, "grad_norm": 0.4240289330482483, "learning_rate": 8e-05, "loss": 1.7707, "step": 6599 }, { "epoch": 0.713282178752837, "grad_norm": 0.43341970443725586, "learning_rate": 8e-05, "loss": 1.6367, "step": 6600 }, { "epoch": 0.7133902518102238, "grad_norm": 0.4390535354614258, "learning_rate": 8e-05, "loss": 1.7223, "step": 6601 }, { "epoch": 0.7134983248676106, "grad_norm": 0.46149998903274536, "learning_rate": 8e-05, "loss": 1.5976, "step": 6602 }, { "epoch": 0.7136063979249972, "grad_norm": 0.42829734086990356, "learning_rate": 8e-05, "loss": 1.6695, "step": 6603 }, { "epoch": 0.713714470982384, "grad_norm": 0.4247375428676605, "learning_rate": 8e-05, "loss": 1.8426, "step": 6604 }, { "epoch": 0.7138225440397709, "grad_norm": 0.4889446198940277, "learning_rate": 8e-05, "loss": 1.8776, "step": 6605 }, { "epoch": 0.7139306170971577, "grad_norm": 0.427287220954895, "learning_rate": 8e-05, "loss": 1.8266, "step": 6606 }, { "epoch": 0.7140386901545445, "grad_norm": 0.460541307926178, "learning_rate": 8e-05, "loss": 1.7166, "step": 6607 }, { "epoch": 0.7141467632119313, "grad_norm": 0.41096630692481995, "learning_rate": 8e-05, "loss": 1.6445, "step": 6608 }, { "epoch": 0.7142548362693181, "grad_norm": 0.4434557557106018, "learning_rate": 8e-05, "loss": 1.8115, "step": 6609 }, { "epoch": 0.7143629093267049, "grad_norm": 0.3965652883052826, "learning_rate": 8e-05, "loss": 1.5474, "step": 6610 }, { "epoch": 0.7144709823840917, "grad_norm": 0.4698472023010254, "learning_rate": 8e-05, "loss": 1.573, "step": 6611 }, { "epoch": 0.7145790554414785, "grad_norm": 0.45088914036750793, "learning_rate": 8e-05, "loss": 1.6348, "step": 6612 }, { "epoch": 0.7146871284988653, "grad_norm": 0.46842294931411743, "learning_rate": 8e-05, "loss": 1.8763, "step": 6613 }, { "epoch": 0.714795201556252, "grad_norm": 0.4356730580329895, "learning_rate": 8e-05, "loss": 1.644, "step": 6614 }, { "epoch": 0.7149032746136388, "grad_norm": 0.475900799036026, "learning_rate": 8e-05, "loss": 1.8466, "step": 6615 }, { "epoch": 0.7150113476710256, "grad_norm": 0.4419645369052887, "learning_rate": 8e-05, "loss": 1.7079, "step": 6616 }, { "epoch": 0.7151194207284124, "grad_norm": 0.4486187696456909, "learning_rate": 8e-05, "loss": 1.6956, "step": 6617 }, { "epoch": 0.7152274937857992, "grad_norm": 0.5011917948722839, "learning_rate": 8e-05, "loss": 1.958, "step": 6618 }, { "epoch": 0.715335566843186, "grad_norm": 0.4425762891769409, "learning_rate": 8e-05, "loss": 1.8077, "step": 6619 }, { "epoch": 0.7154436399005728, "grad_norm": 0.48137739300727844, "learning_rate": 8e-05, "loss": 1.7533, "step": 6620 }, { "epoch": 0.7155517129579596, "grad_norm": 0.4081377685070038, "learning_rate": 8e-05, "loss": 1.5716, "step": 6621 }, { "epoch": 0.7156597860153464, "grad_norm": 0.4411361813545227, "learning_rate": 8e-05, "loss": 1.794, "step": 6622 }, { "epoch": 0.7157678590727332, "grad_norm": 0.426044225692749, "learning_rate": 8e-05, "loss": 1.7238, "step": 6623 }, { "epoch": 0.7158759321301199, "grad_norm": 0.43416014313697815, "learning_rate": 8e-05, "loss": 1.5876, "step": 6624 }, { "epoch": 0.7159840051875067, "grad_norm": 0.4871584475040436, "learning_rate": 8e-05, "loss": 1.7176, "step": 6625 }, { "epoch": 0.7160920782448935, "grad_norm": 0.4505137801170349, "learning_rate": 8e-05, "loss": 1.4732, "step": 6626 }, { "epoch": 0.7162001513022803, "grad_norm": 0.4161204695701599, "learning_rate": 8e-05, "loss": 1.5161, "step": 6627 }, { "epoch": 0.7163082243596671, "grad_norm": 0.43388932943344116, "learning_rate": 8e-05, "loss": 1.4509, "step": 6628 }, { "epoch": 0.7164162974170539, "grad_norm": 0.4667082726955414, "learning_rate": 8e-05, "loss": 1.6952, "step": 6629 }, { "epoch": 0.7165243704744407, "grad_norm": 0.5074205994606018, "learning_rate": 8e-05, "loss": 1.7343, "step": 6630 }, { "epoch": 0.7166324435318275, "grad_norm": 0.4615396559238434, "learning_rate": 8e-05, "loss": 1.7228, "step": 6631 }, { "epoch": 0.7167405165892143, "grad_norm": 0.4753437638282776, "learning_rate": 8e-05, "loss": 1.6414, "step": 6632 }, { "epoch": 0.7168485896466011, "grad_norm": 0.3991701006889343, "learning_rate": 8e-05, "loss": 1.5658, "step": 6633 }, { "epoch": 0.716956662703988, "grad_norm": 0.44477176666259766, "learning_rate": 8e-05, "loss": 1.7328, "step": 6634 }, { "epoch": 0.7170647357613746, "grad_norm": 0.4382745623588562, "learning_rate": 8e-05, "loss": 1.7033, "step": 6635 }, { "epoch": 0.7171728088187614, "grad_norm": 0.39922448992729187, "learning_rate": 8e-05, "loss": 1.6649, "step": 6636 }, { "epoch": 0.7172808818761482, "grad_norm": 0.4345834255218506, "learning_rate": 8e-05, "loss": 1.4544, "step": 6637 }, { "epoch": 0.717388954933535, "grad_norm": 0.43980902433395386, "learning_rate": 8e-05, "loss": 1.5833, "step": 6638 }, { "epoch": 0.7174970279909219, "grad_norm": 0.46017521619796753, "learning_rate": 8e-05, "loss": 1.7454, "step": 6639 }, { "epoch": 0.7176051010483087, "grad_norm": 0.4095003008842468, "learning_rate": 8e-05, "loss": 1.6226, "step": 6640 }, { "epoch": 0.7177131741056955, "grad_norm": 0.4101576805114746, "learning_rate": 8e-05, "loss": 1.6885, "step": 6641 }, { "epoch": 0.7178212471630823, "grad_norm": 0.4970894157886505, "learning_rate": 8e-05, "loss": 1.9697, "step": 6642 }, { "epoch": 0.7179293202204691, "grad_norm": 0.42190656065940857, "learning_rate": 8e-05, "loss": 1.6507, "step": 6643 }, { "epoch": 0.7180373932778559, "grad_norm": 0.40890735387802124, "learning_rate": 8e-05, "loss": 1.5924, "step": 6644 }, { "epoch": 0.7181454663352427, "grad_norm": 0.4323946237564087, "learning_rate": 8e-05, "loss": 1.5859, "step": 6645 }, { "epoch": 0.7182535393926294, "grad_norm": 0.41717079281806946, "learning_rate": 8e-05, "loss": 1.7223, "step": 6646 }, { "epoch": 0.7183616124500162, "grad_norm": 0.41902562975883484, "learning_rate": 8e-05, "loss": 1.7017, "step": 6647 }, { "epoch": 0.718469685507403, "grad_norm": 0.48267847299575806, "learning_rate": 8e-05, "loss": 1.7214, "step": 6648 }, { "epoch": 0.7185777585647898, "grad_norm": 0.4220077395439148, "learning_rate": 8e-05, "loss": 1.6856, "step": 6649 }, { "epoch": 0.7186858316221766, "grad_norm": 0.4507480263710022, "learning_rate": 8e-05, "loss": 1.8113, "step": 6650 }, { "epoch": 0.7187939046795634, "grad_norm": 0.44540709257125854, "learning_rate": 8e-05, "loss": 1.6283, "step": 6651 }, { "epoch": 0.7189019777369502, "grad_norm": 0.4751279950141907, "learning_rate": 8e-05, "loss": 1.792, "step": 6652 }, { "epoch": 0.719010050794337, "grad_norm": 0.40676558017730713, "learning_rate": 8e-05, "loss": 1.6455, "step": 6653 }, { "epoch": 0.7191181238517238, "grad_norm": 0.4524264633655548, "learning_rate": 8e-05, "loss": 1.7793, "step": 6654 }, { "epoch": 0.7192261969091106, "grad_norm": 0.4806995987892151, "learning_rate": 8e-05, "loss": 1.7792, "step": 6655 }, { "epoch": 0.7193342699664973, "grad_norm": 0.4639224708080292, "learning_rate": 8e-05, "loss": 1.8478, "step": 6656 }, { "epoch": 0.7194423430238841, "grad_norm": 0.4594043493270874, "learning_rate": 8e-05, "loss": 1.7125, "step": 6657 }, { "epoch": 0.7195504160812709, "grad_norm": 0.4194337725639343, "learning_rate": 8e-05, "loss": 1.6531, "step": 6658 }, { "epoch": 0.7196584891386577, "grad_norm": 0.47044244408607483, "learning_rate": 8e-05, "loss": 1.786, "step": 6659 }, { "epoch": 0.7197665621960445, "grad_norm": 0.4315778315067291, "learning_rate": 8e-05, "loss": 1.6904, "step": 6660 }, { "epoch": 0.7198746352534313, "grad_norm": 0.43112415075302124, "learning_rate": 8e-05, "loss": 1.7392, "step": 6661 }, { "epoch": 0.7199827083108181, "grad_norm": 0.5010254979133606, "learning_rate": 8e-05, "loss": 1.9478, "step": 6662 }, { "epoch": 0.7200907813682049, "grad_norm": 0.434428334236145, "learning_rate": 8e-05, "loss": 1.6169, "step": 6663 }, { "epoch": 0.7201988544255917, "grad_norm": 0.4186517298221588, "learning_rate": 8e-05, "loss": 1.8174, "step": 6664 }, { "epoch": 0.7203069274829785, "grad_norm": 0.41959133744239807, "learning_rate": 8e-05, "loss": 1.6138, "step": 6665 }, { "epoch": 0.7204150005403653, "grad_norm": 0.42690449953079224, "learning_rate": 8e-05, "loss": 1.6776, "step": 6666 }, { "epoch": 0.720523073597752, "grad_norm": 0.4610652029514313, "learning_rate": 8e-05, "loss": 1.7829, "step": 6667 }, { "epoch": 0.7206311466551388, "grad_norm": 0.4553520381450653, "learning_rate": 8e-05, "loss": 1.6254, "step": 6668 }, { "epoch": 0.7207392197125256, "grad_norm": 0.4146069586277008, "learning_rate": 8e-05, "loss": 1.6014, "step": 6669 }, { "epoch": 0.7208472927699124, "grad_norm": 0.44331613183021545, "learning_rate": 8e-05, "loss": 1.7406, "step": 6670 }, { "epoch": 0.7209553658272992, "grad_norm": 0.4194159209728241, "learning_rate": 8e-05, "loss": 1.8009, "step": 6671 }, { "epoch": 0.721063438884686, "grad_norm": 0.4974169135093689, "learning_rate": 8e-05, "loss": 1.7226, "step": 6672 }, { "epoch": 0.7211715119420729, "grad_norm": 0.41169628500938416, "learning_rate": 8e-05, "loss": 1.6803, "step": 6673 }, { "epoch": 0.7212795849994597, "grad_norm": 0.39743897318840027, "learning_rate": 8e-05, "loss": 1.5711, "step": 6674 }, { "epoch": 0.7213876580568465, "grad_norm": 0.4424927830696106, "learning_rate": 8e-05, "loss": 1.7685, "step": 6675 }, { "epoch": 0.7214957311142333, "grad_norm": 0.4413016736507416, "learning_rate": 8e-05, "loss": 1.8475, "step": 6676 }, { "epoch": 0.7216038041716201, "grad_norm": 0.44189026951789856, "learning_rate": 8e-05, "loss": 1.659, "step": 6677 }, { "epoch": 0.7217118772290068, "grad_norm": 0.4550066590309143, "learning_rate": 8e-05, "loss": 1.6356, "step": 6678 }, { "epoch": 0.7218199502863936, "grad_norm": 0.43385645747184753, "learning_rate": 8e-05, "loss": 1.6652, "step": 6679 }, { "epoch": 0.7219280233437804, "grad_norm": 0.45742782950401306, "learning_rate": 8e-05, "loss": 1.7771, "step": 6680 }, { "epoch": 0.7220360964011672, "grad_norm": 0.42534756660461426, "learning_rate": 8e-05, "loss": 1.6347, "step": 6681 }, { "epoch": 0.722144169458554, "grad_norm": 0.4059029817581177, "learning_rate": 8e-05, "loss": 1.5337, "step": 6682 }, { "epoch": 0.7222522425159408, "grad_norm": 0.44792819023132324, "learning_rate": 8e-05, "loss": 1.7756, "step": 6683 }, { "epoch": 0.7223603155733276, "grad_norm": 0.46006935834884644, "learning_rate": 8e-05, "loss": 1.8537, "step": 6684 }, { "epoch": 0.7224683886307144, "grad_norm": 0.43359193205833435, "learning_rate": 8e-05, "loss": 1.7497, "step": 6685 }, { "epoch": 0.7225764616881012, "grad_norm": 0.43468308448791504, "learning_rate": 8e-05, "loss": 1.7767, "step": 6686 }, { "epoch": 0.722684534745488, "grad_norm": 0.44941258430480957, "learning_rate": 8e-05, "loss": 1.815, "step": 6687 }, { "epoch": 0.7227926078028748, "grad_norm": 0.48379504680633545, "learning_rate": 8e-05, "loss": 1.7903, "step": 6688 }, { "epoch": 0.7229006808602615, "grad_norm": 0.4190543293952942, "learning_rate": 8e-05, "loss": 1.695, "step": 6689 }, { "epoch": 0.7230087539176483, "grad_norm": 0.40381571650505066, "learning_rate": 8e-05, "loss": 1.5357, "step": 6690 }, { "epoch": 0.7231168269750351, "grad_norm": 0.4146292507648468, "learning_rate": 8e-05, "loss": 1.6767, "step": 6691 }, { "epoch": 0.7232249000324219, "grad_norm": 0.44261685013771057, "learning_rate": 8e-05, "loss": 1.642, "step": 6692 }, { "epoch": 0.7233329730898087, "grad_norm": 0.5104547739028931, "learning_rate": 8e-05, "loss": 1.7571, "step": 6693 }, { "epoch": 0.7234410461471955, "grad_norm": 0.43287134170532227, "learning_rate": 8e-05, "loss": 1.807, "step": 6694 }, { "epoch": 0.7235491192045823, "grad_norm": 0.4997626841068268, "learning_rate": 8e-05, "loss": 1.9568, "step": 6695 }, { "epoch": 0.7236571922619691, "grad_norm": 0.40999653935432434, "learning_rate": 8e-05, "loss": 1.6001, "step": 6696 }, { "epoch": 0.7237652653193559, "grad_norm": 0.43269169330596924, "learning_rate": 8e-05, "loss": 1.7295, "step": 6697 }, { "epoch": 0.7238733383767427, "grad_norm": 0.42358776926994324, "learning_rate": 8e-05, "loss": 1.7158, "step": 6698 }, { "epoch": 0.7239814114341294, "grad_norm": 0.3992778956890106, "learning_rate": 8e-05, "loss": 1.6107, "step": 6699 }, { "epoch": 0.7240894844915162, "grad_norm": 0.4853735566139221, "learning_rate": 8e-05, "loss": 1.7521, "step": 6700 }, { "epoch": 0.724197557548903, "grad_norm": 0.41710156202316284, "learning_rate": 8e-05, "loss": 1.6709, "step": 6701 }, { "epoch": 0.7243056306062898, "grad_norm": 0.395452618598938, "learning_rate": 8e-05, "loss": 1.6567, "step": 6702 }, { "epoch": 0.7244137036636766, "grad_norm": 0.419277161359787, "learning_rate": 8e-05, "loss": 1.7142, "step": 6703 }, { "epoch": 0.7245217767210634, "grad_norm": 0.41060853004455566, "learning_rate": 8e-05, "loss": 1.5132, "step": 6704 }, { "epoch": 0.7246298497784502, "grad_norm": 0.4331308901309967, "learning_rate": 8e-05, "loss": 1.7151, "step": 6705 }, { "epoch": 0.724737922835837, "grad_norm": 0.441701203584671, "learning_rate": 8e-05, "loss": 1.732, "step": 6706 }, { "epoch": 0.7248459958932238, "grad_norm": 0.4894917905330658, "learning_rate": 8e-05, "loss": 1.7645, "step": 6707 }, { "epoch": 0.7249540689506107, "grad_norm": 0.40846335887908936, "learning_rate": 8e-05, "loss": 1.5882, "step": 6708 }, { "epoch": 0.7250621420079975, "grad_norm": 0.40337637066841125, "learning_rate": 8e-05, "loss": 1.5795, "step": 6709 }, { "epoch": 0.7251702150653841, "grad_norm": 0.41570696234703064, "learning_rate": 8e-05, "loss": 1.6441, "step": 6710 }, { "epoch": 0.725278288122771, "grad_norm": 0.41095322370529175, "learning_rate": 8e-05, "loss": 1.6976, "step": 6711 }, { "epoch": 0.7253863611801578, "grad_norm": 0.4408228099346161, "learning_rate": 8e-05, "loss": 1.7369, "step": 6712 }, { "epoch": 0.7254944342375446, "grad_norm": 0.45103904604911804, "learning_rate": 8e-05, "loss": 1.7198, "step": 6713 }, { "epoch": 0.7256025072949314, "grad_norm": 0.4298135042190552, "learning_rate": 8e-05, "loss": 1.7696, "step": 6714 }, { "epoch": 0.7257105803523182, "grad_norm": 0.40494397282600403, "learning_rate": 8e-05, "loss": 1.5304, "step": 6715 }, { "epoch": 0.725818653409705, "grad_norm": 0.4021197259426117, "learning_rate": 8e-05, "loss": 1.7551, "step": 6716 }, { "epoch": 0.7259267264670918, "grad_norm": 0.4598579406738281, "learning_rate": 8e-05, "loss": 1.8588, "step": 6717 }, { "epoch": 0.7260347995244786, "grad_norm": 0.41416507959365845, "learning_rate": 8e-05, "loss": 1.652, "step": 6718 }, { "epoch": 0.7261428725818654, "grad_norm": 0.4621504843235016, "learning_rate": 8e-05, "loss": 1.777, "step": 6719 }, { "epoch": 0.7262509456392522, "grad_norm": 0.4476461112499237, "learning_rate": 8e-05, "loss": 1.7217, "step": 6720 }, { "epoch": 0.7263590186966389, "grad_norm": 0.4339551627635956, "learning_rate": 8e-05, "loss": 1.6235, "step": 6721 }, { "epoch": 0.7264670917540257, "grad_norm": 0.44488951563835144, "learning_rate": 8e-05, "loss": 1.7139, "step": 6722 }, { "epoch": 0.7265751648114125, "grad_norm": 0.4391340911388397, "learning_rate": 8e-05, "loss": 1.7848, "step": 6723 }, { "epoch": 0.7266832378687993, "grad_norm": 0.40188461542129517, "learning_rate": 8e-05, "loss": 1.7255, "step": 6724 }, { "epoch": 0.7267913109261861, "grad_norm": 0.43013763427734375, "learning_rate": 8e-05, "loss": 1.6746, "step": 6725 }, { "epoch": 0.7268993839835729, "grad_norm": 0.442081481218338, "learning_rate": 8e-05, "loss": 1.7425, "step": 6726 }, { "epoch": 0.7270074570409597, "grad_norm": 0.42574265599250793, "learning_rate": 8e-05, "loss": 1.5829, "step": 6727 }, { "epoch": 0.7271155300983465, "grad_norm": 0.43970537185668945, "learning_rate": 8e-05, "loss": 1.7376, "step": 6728 }, { "epoch": 0.7272236031557333, "grad_norm": 0.39443281292915344, "learning_rate": 8e-05, "loss": 1.6684, "step": 6729 }, { "epoch": 0.7273316762131201, "grad_norm": 0.4474886357784271, "learning_rate": 8e-05, "loss": 1.8383, "step": 6730 }, { "epoch": 0.7274397492705068, "grad_norm": 0.42771419882774353, "learning_rate": 8e-05, "loss": 1.645, "step": 6731 }, { "epoch": 0.7275478223278936, "grad_norm": 0.45707783102989197, "learning_rate": 8e-05, "loss": 1.8683, "step": 6732 }, { "epoch": 0.7276558953852804, "grad_norm": 0.43292000889778137, "learning_rate": 8e-05, "loss": 1.6593, "step": 6733 }, { "epoch": 0.7277639684426672, "grad_norm": 0.42142125964164734, "learning_rate": 8e-05, "loss": 1.6834, "step": 6734 }, { "epoch": 0.727872041500054, "grad_norm": 0.41964399814605713, "learning_rate": 8e-05, "loss": 1.7054, "step": 6735 }, { "epoch": 0.7279801145574408, "grad_norm": 0.4104583263397217, "learning_rate": 8e-05, "loss": 1.5454, "step": 6736 }, { "epoch": 0.7280881876148276, "grad_norm": 0.4196583330631256, "learning_rate": 8e-05, "loss": 1.6403, "step": 6737 }, { "epoch": 0.7281962606722144, "grad_norm": 0.495767742395401, "learning_rate": 8e-05, "loss": 1.737, "step": 6738 }, { "epoch": 0.7283043337296012, "grad_norm": 0.42700427770614624, "learning_rate": 8e-05, "loss": 1.6436, "step": 6739 }, { "epoch": 0.728412406786988, "grad_norm": 0.46040868759155273, "learning_rate": 8e-05, "loss": 1.7028, "step": 6740 }, { "epoch": 0.7285204798443748, "grad_norm": 0.4147094488143921, "learning_rate": 8e-05, "loss": 1.7214, "step": 6741 }, { "epoch": 0.7286285529017615, "grad_norm": 0.4350791573524475, "learning_rate": 8e-05, "loss": 1.6168, "step": 6742 }, { "epoch": 0.7287366259591483, "grad_norm": 0.39815229177474976, "learning_rate": 8e-05, "loss": 1.6117, "step": 6743 }, { "epoch": 0.7288446990165351, "grad_norm": 0.48351773619651794, "learning_rate": 8e-05, "loss": 1.701, "step": 6744 }, { "epoch": 0.728952772073922, "grad_norm": 0.3839779794216156, "learning_rate": 8e-05, "loss": 1.4266, "step": 6745 }, { "epoch": 0.7290608451313088, "grad_norm": 0.44250375032424927, "learning_rate": 8e-05, "loss": 1.4355, "step": 6746 }, { "epoch": 0.7291689181886956, "grad_norm": 0.4060174226760864, "learning_rate": 8e-05, "loss": 1.6211, "step": 6747 }, { "epoch": 0.7292769912460824, "grad_norm": 0.517372727394104, "learning_rate": 8e-05, "loss": 1.9127, "step": 6748 }, { "epoch": 0.7293850643034692, "grad_norm": 0.41998621821403503, "learning_rate": 8e-05, "loss": 1.5611, "step": 6749 }, { "epoch": 0.729493137360856, "grad_norm": 0.3964490294456482, "learning_rate": 8e-05, "loss": 1.5608, "step": 6750 }, { "epoch": 0.7296012104182428, "grad_norm": 0.47403374314308167, "learning_rate": 8e-05, "loss": 1.7509, "step": 6751 }, { "epoch": 0.7297092834756296, "grad_norm": 0.4228605628013611, "learning_rate": 8e-05, "loss": 1.609, "step": 6752 }, { "epoch": 0.7298173565330163, "grad_norm": 0.4627263844013214, "learning_rate": 8e-05, "loss": 1.7454, "step": 6753 }, { "epoch": 0.7299254295904031, "grad_norm": 0.4039344787597656, "learning_rate": 8e-05, "loss": 1.5901, "step": 6754 }, { "epoch": 0.7300335026477899, "grad_norm": 0.46222615242004395, "learning_rate": 8e-05, "loss": 1.8116, "step": 6755 }, { "epoch": 0.7301415757051767, "grad_norm": 0.4364181458950043, "learning_rate": 8e-05, "loss": 1.7274, "step": 6756 }, { "epoch": 0.7302496487625635, "grad_norm": 0.470706045627594, "learning_rate": 8e-05, "loss": 1.8199, "step": 6757 }, { "epoch": 0.7303577218199503, "grad_norm": 0.41992005705833435, "learning_rate": 8e-05, "loss": 1.414, "step": 6758 }, { "epoch": 0.7304657948773371, "grad_norm": 0.4663428068161011, "learning_rate": 8e-05, "loss": 1.845, "step": 6759 }, { "epoch": 0.7305738679347239, "grad_norm": 0.42762741446495056, "learning_rate": 8e-05, "loss": 1.7367, "step": 6760 }, { "epoch": 0.7306819409921107, "grad_norm": 0.4279184937477112, "learning_rate": 8e-05, "loss": 1.7659, "step": 6761 }, { "epoch": 0.7307900140494975, "grad_norm": 0.45592308044433594, "learning_rate": 8e-05, "loss": 1.8521, "step": 6762 }, { "epoch": 0.7308980871068843, "grad_norm": 0.4446328282356262, "learning_rate": 8e-05, "loss": 1.7574, "step": 6763 }, { "epoch": 0.731006160164271, "grad_norm": 0.4486009180545807, "learning_rate": 8e-05, "loss": 1.6647, "step": 6764 }, { "epoch": 0.7311142332216578, "grad_norm": 0.49798640608787537, "learning_rate": 8e-05, "loss": 1.8825, "step": 6765 }, { "epoch": 0.7312223062790446, "grad_norm": 0.4332321584224701, "learning_rate": 8e-05, "loss": 1.6715, "step": 6766 }, { "epoch": 0.7313303793364314, "grad_norm": 0.42355623841285706, "learning_rate": 8e-05, "loss": 1.6558, "step": 6767 }, { "epoch": 0.7314384523938182, "grad_norm": 0.4275186359882355, "learning_rate": 8e-05, "loss": 1.7974, "step": 6768 }, { "epoch": 0.731546525451205, "grad_norm": 0.4250764548778534, "learning_rate": 8e-05, "loss": 1.5326, "step": 6769 }, { "epoch": 0.7316545985085918, "grad_norm": 0.4315224289894104, "learning_rate": 8e-05, "loss": 1.6198, "step": 6770 }, { "epoch": 0.7317626715659786, "grad_norm": 0.4442756772041321, "learning_rate": 8e-05, "loss": 1.6839, "step": 6771 }, { "epoch": 0.7318707446233654, "grad_norm": 0.4185662865638733, "learning_rate": 8e-05, "loss": 1.562, "step": 6772 }, { "epoch": 0.7319788176807522, "grad_norm": 0.4162086248397827, "learning_rate": 8e-05, "loss": 1.609, "step": 6773 }, { "epoch": 0.7320868907381389, "grad_norm": 0.441159725189209, "learning_rate": 8e-05, "loss": 1.7346, "step": 6774 }, { "epoch": 0.7321949637955257, "grad_norm": 0.42583996057510376, "learning_rate": 8e-05, "loss": 1.6753, "step": 6775 }, { "epoch": 0.7323030368529125, "grad_norm": 0.4658140242099762, "learning_rate": 8e-05, "loss": 1.6649, "step": 6776 }, { "epoch": 0.7324111099102993, "grad_norm": 0.4370420277118683, "learning_rate": 8e-05, "loss": 1.6413, "step": 6777 }, { "epoch": 0.7325191829676861, "grad_norm": 0.420751690864563, "learning_rate": 8e-05, "loss": 1.6545, "step": 6778 }, { "epoch": 0.732627256025073, "grad_norm": 0.42113232612609863, "learning_rate": 8e-05, "loss": 1.6625, "step": 6779 }, { "epoch": 0.7327353290824598, "grad_norm": 0.4502719044685364, "learning_rate": 8e-05, "loss": 1.6995, "step": 6780 }, { "epoch": 0.7328434021398466, "grad_norm": 0.42127272486686707, "learning_rate": 8e-05, "loss": 1.5696, "step": 6781 }, { "epoch": 0.7329514751972334, "grad_norm": 0.4633394479751587, "learning_rate": 8e-05, "loss": 1.5076, "step": 6782 }, { "epoch": 0.7330595482546202, "grad_norm": 0.46363359689712524, "learning_rate": 8e-05, "loss": 1.7347, "step": 6783 }, { "epoch": 0.733167621312007, "grad_norm": 0.48651283979415894, "learning_rate": 8e-05, "loss": 1.5745, "step": 6784 }, { "epoch": 0.7332756943693937, "grad_norm": 0.47666627168655396, "learning_rate": 8e-05, "loss": 1.7533, "step": 6785 }, { "epoch": 0.7333837674267805, "grad_norm": 0.4426896870136261, "learning_rate": 8e-05, "loss": 1.6318, "step": 6786 }, { "epoch": 0.7334918404841673, "grad_norm": 0.44029054045677185, "learning_rate": 8e-05, "loss": 1.5693, "step": 6787 }, { "epoch": 0.7335999135415541, "grad_norm": 0.4171055555343628, "learning_rate": 8e-05, "loss": 1.663, "step": 6788 }, { "epoch": 0.7337079865989409, "grad_norm": 0.4296603500843048, "learning_rate": 8e-05, "loss": 1.6541, "step": 6789 }, { "epoch": 0.7338160596563277, "grad_norm": 1.2785691022872925, "learning_rate": 8e-05, "loss": 1.7944, "step": 6790 }, { "epoch": 0.7339241327137145, "grad_norm": 0.4807487726211548, "learning_rate": 8e-05, "loss": 1.6742, "step": 6791 }, { "epoch": 0.7340322057711013, "grad_norm": 0.49236124753952026, "learning_rate": 8e-05, "loss": 1.8218, "step": 6792 }, { "epoch": 0.7341402788284881, "grad_norm": 0.42554065585136414, "learning_rate": 8e-05, "loss": 1.5159, "step": 6793 }, { "epoch": 0.7342483518858749, "grad_norm": 0.49967601895332336, "learning_rate": 8e-05, "loss": 1.8977, "step": 6794 }, { "epoch": 0.7343564249432617, "grad_norm": 0.4496057629585266, "learning_rate": 8e-05, "loss": 1.8115, "step": 6795 }, { "epoch": 0.7344644980006484, "grad_norm": 0.41880321502685547, "learning_rate": 8e-05, "loss": 1.7192, "step": 6796 }, { "epoch": 0.7345725710580352, "grad_norm": 0.4451134204864502, "learning_rate": 8e-05, "loss": 1.735, "step": 6797 }, { "epoch": 0.734680644115422, "grad_norm": 0.43343856930732727, "learning_rate": 8e-05, "loss": 1.6584, "step": 6798 }, { "epoch": 0.7347887171728088, "grad_norm": 0.424786776304245, "learning_rate": 8e-05, "loss": 1.7722, "step": 6799 }, { "epoch": 0.7348967902301956, "grad_norm": 0.4430929124355316, "learning_rate": 8e-05, "loss": 1.6608, "step": 6800 }, { "epoch": 0.7350048632875824, "grad_norm": 0.4264802038669586, "learning_rate": 8e-05, "loss": 1.6624, "step": 6801 }, { "epoch": 0.7351129363449692, "grad_norm": 0.39704233407974243, "learning_rate": 8e-05, "loss": 1.4464, "step": 6802 }, { "epoch": 0.735221009402356, "grad_norm": 0.4317352771759033, "learning_rate": 8e-05, "loss": 1.7323, "step": 6803 }, { "epoch": 0.7353290824597428, "grad_norm": 0.4472956359386444, "learning_rate": 8e-05, "loss": 1.7371, "step": 6804 }, { "epoch": 0.7354371555171296, "grad_norm": 0.4822424054145813, "learning_rate": 8e-05, "loss": 1.6741, "step": 6805 }, { "epoch": 0.7355452285745163, "grad_norm": 0.41965577006340027, "learning_rate": 8e-05, "loss": 1.4825, "step": 6806 }, { "epoch": 0.7356533016319031, "grad_norm": 0.4579613208770752, "learning_rate": 8e-05, "loss": 1.817, "step": 6807 }, { "epoch": 0.7357613746892899, "grad_norm": 0.44924747943878174, "learning_rate": 8e-05, "loss": 1.7468, "step": 6808 }, { "epoch": 0.7358694477466767, "grad_norm": 0.425653338432312, "learning_rate": 8e-05, "loss": 1.6574, "step": 6809 }, { "epoch": 0.7359775208040635, "grad_norm": 0.44882914423942566, "learning_rate": 8e-05, "loss": 1.66, "step": 6810 }, { "epoch": 0.7360855938614503, "grad_norm": 0.43751975893974304, "learning_rate": 8e-05, "loss": 1.6646, "step": 6811 }, { "epoch": 0.7361936669188371, "grad_norm": 0.43542608618736267, "learning_rate": 8e-05, "loss": 1.7326, "step": 6812 }, { "epoch": 0.736301739976224, "grad_norm": 0.42972058057785034, "learning_rate": 8e-05, "loss": 1.7339, "step": 6813 }, { "epoch": 0.7364098130336107, "grad_norm": 0.41918709874153137, "learning_rate": 8e-05, "loss": 1.7349, "step": 6814 }, { "epoch": 0.7365178860909976, "grad_norm": 0.4454612731933594, "learning_rate": 8e-05, "loss": 1.7639, "step": 6815 }, { "epoch": 0.7366259591483844, "grad_norm": 0.396497905254364, "learning_rate": 8e-05, "loss": 1.6073, "step": 6816 }, { "epoch": 0.736734032205771, "grad_norm": 0.4799298644065857, "learning_rate": 8e-05, "loss": 1.7422, "step": 6817 }, { "epoch": 0.7368421052631579, "grad_norm": 0.4155880808830261, "learning_rate": 8e-05, "loss": 1.7237, "step": 6818 }, { "epoch": 0.7369501783205447, "grad_norm": 0.4390081465244293, "learning_rate": 8e-05, "loss": 1.6737, "step": 6819 }, { "epoch": 0.7370582513779315, "grad_norm": 0.40610557794570923, "learning_rate": 8e-05, "loss": 1.6284, "step": 6820 }, { "epoch": 0.7371663244353183, "grad_norm": 0.409546822309494, "learning_rate": 8e-05, "loss": 1.6217, "step": 6821 }, { "epoch": 0.7372743974927051, "grad_norm": 0.44896286725997925, "learning_rate": 8e-05, "loss": 1.4609, "step": 6822 }, { "epoch": 0.7373824705500919, "grad_norm": 0.48128920793533325, "learning_rate": 8e-05, "loss": 1.6852, "step": 6823 }, { "epoch": 0.7374905436074787, "grad_norm": 0.41410893201828003, "learning_rate": 8e-05, "loss": 1.6098, "step": 6824 }, { "epoch": 0.7375986166648655, "grad_norm": 0.43301814794540405, "learning_rate": 8e-05, "loss": 1.6392, "step": 6825 }, { "epoch": 0.7377066897222523, "grad_norm": 0.45773082971572876, "learning_rate": 8e-05, "loss": 1.7544, "step": 6826 }, { "epoch": 0.7378147627796391, "grad_norm": 0.40748131275177, "learning_rate": 8e-05, "loss": 1.673, "step": 6827 }, { "epoch": 0.7379228358370258, "grad_norm": 0.4408351480960846, "learning_rate": 8e-05, "loss": 1.5956, "step": 6828 }, { "epoch": 0.7380309088944126, "grad_norm": 0.4522993862628937, "learning_rate": 8e-05, "loss": 1.5547, "step": 6829 }, { "epoch": 0.7381389819517994, "grad_norm": 0.40294572710990906, "learning_rate": 8e-05, "loss": 1.6247, "step": 6830 }, { "epoch": 0.7382470550091862, "grad_norm": 0.4294702708721161, "learning_rate": 8e-05, "loss": 1.8537, "step": 6831 }, { "epoch": 0.738355128066573, "grad_norm": 0.42691946029663086, "learning_rate": 8e-05, "loss": 1.7131, "step": 6832 }, { "epoch": 0.7384632011239598, "grad_norm": 0.42257192730903625, "learning_rate": 8e-05, "loss": 1.563, "step": 6833 }, { "epoch": 0.7385712741813466, "grad_norm": 0.4658922553062439, "learning_rate": 8e-05, "loss": 1.8242, "step": 6834 }, { "epoch": 0.7386793472387334, "grad_norm": 0.40969598293304443, "learning_rate": 8e-05, "loss": 1.735, "step": 6835 }, { "epoch": 0.7387874202961202, "grad_norm": 0.41168689727783203, "learning_rate": 8e-05, "loss": 1.5856, "step": 6836 }, { "epoch": 0.738895493353507, "grad_norm": 0.44230201840400696, "learning_rate": 8e-05, "loss": 1.7398, "step": 6837 }, { "epoch": 0.7390035664108938, "grad_norm": 0.47442471981048584, "learning_rate": 8e-05, "loss": 1.8702, "step": 6838 }, { "epoch": 0.7391116394682805, "grad_norm": 0.4787534177303314, "learning_rate": 8e-05, "loss": 1.803, "step": 6839 }, { "epoch": 0.7392197125256673, "grad_norm": 0.4366009831428528, "learning_rate": 8e-05, "loss": 1.768, "step": 6840 }, { "epoch": 0.7393277855830541, "grad_norm": 0.4298141896724701, "learning_rate": 8e-05, "loss": 1.787, "step": 6841 }, { "epoch": 0.7394358586404409, "grad_norm": 0.424793004989624, "learning_rate": 8e-05, "loss": 1.6645, "step": 6842 }, { "epoch": 0.7395439316978277, "grad_norm": 0.4245516061782837, "learning_rate": 8e-05, "loss": 1.6087, "step": 6843 }, { "epoch": 0.7396520047552145, "grad_norm": 0.4444918930530548, "learning_rate": 8e-05, "loss": 1.5688, "step": 6844 }, { "epoch": 0.7397600778126013, "grad_norm": 0.4692555069923401, "learning_rate": 8e-05, "loss": 1.9288, "step": 6845 }, { "epoch": 0.7398681508699881, "grad_norm": 0.43688860535621643, "learning_rate": 8e-05, "loss": 1.7065, "step": 6846 }, { "epoch": 0.7399762239273749, "grad_norm": 0.4163924753665924, "learning_rate": 8e-05, "loss": 1.662, "step": 6847 }, { "epoch": 0.7400842969847617, "grad_norm": 0.5075752139091492, "learning_rate": 8e-05, "loss": 1.8612, "step": 6848 }, { "epoch": 0.7401923700421484, "grad_norm": 0.4274742603302002, "learning_rate": 8e-05, "loss": 1.7231, "step": 6849 }, { "epoch": 0.7403004430995352, "grad_norm": 0.45574870705604553, "learning_rate": 8e-05, "loss": 1.6391, "step": 6850 }, { "epoch": 0.740408516156922, "grad_norm": 0.41290149092674255, "learning_rate": 8e-05, "loss": 1.7235, "step": 6851 }, { "epoch": 0.7405165892143089, "grad_norm": 0.46124711632728577, "learning_rate": 8e-05, "loss": 1.8127, "step": 6852 }, { "epoch": 0.7406246622716957, "grad_norm": 0.42965665459632874, "learning_rate": 8e-05, "loss": 1.7913, "step": 6853 }, { "epoch": 0.7407327353290825, "grad_norm": 0.44877204298973083, "learning_rate": 8e-05, "loss": 1.8293, "step": 6854 }, { "epoch": 0.7408408083864693, "grad_norm": 0.4206621050834656, "learning_rate": 8e-05, "loss": 1.6033, "step": 6855 }, { "epoch": 0.7409488814438561, "grad_norm": 0.4061332046985626, "learning_rate": 8e-05, "loss": 1.5834, "step": 6856 }, { "epoch": 0.7410569545012429, "grad_norm": 0.4304428696632385, "learning_rate": 8e-05, "loss": 1.7096, "step": 6857 }, { "epoch": 0.7411650275586297, "grad_norm": 0.47459661960601807, "learning_rate": 8e-05, "loss": 1.7406, "step": 6858 }, { "epoch": 0.7412731006160165, "grad_norm": 0.4122685194015503, "learning_rate": 8e-05, "loss": 1.6294, "step": 6859 }, { "epoch": 0.7413811736734032, "grad_norm": 0.45497190952301025, "learning_rate": 8e-05, "loss": 1.7005, "step": 6860 }, { "epoch": 0.74148924673079, "grad_norm": 0.4115236699581146, "learning_rate": 8e-05, "loss": 1.4768, "step": 6861 }, { "epoch": 0.7415973197881768, "grad_norm": 0.4381580054759979, "learning_rate": 8e-05, "loss": 1.7736, "step": 6862 }, { "epoch": 0.7417053928455636, "grad_norm": 0.4303855895996094, "learning_rate": 8e-05, "loss": 1.6214, "step": 6863 }, { "epoch": 0.7418134659029504, "grad_norm": 0.46382448077201843, "learning_rate": 8e-05, "loss": 1.9042, "step": 6864 }, { "epoch": 0.7419215389603372, "grad_norm": 0.44578033685684204, "learning_rate": 8e-05, "loss": 1.6201, "step": 6865 }, { "epoch": 0.742029612017724, "grad_norm": 0.42521071434020996, "learning_rate": 8e-05, "loss": 1.7253, "step": 6866 }, { "epoch": 0.7421376850751108, "grad_norm": 0.4140975773334503, "learning_rate": 8e-05, "loss": 1.5683, "step": 6867 }, { "epoch": 0.7422457581324976, "grad_norm": 0.43179816007614136, "learning_rate": 8e-05, "loss": 1.5785, "step": 6868 }, { "epoch": 0.7423538311898844, "grad_norm": 0.42436152696609497, "learning_rate": 8e-05, "loss": 1.5245, "step": 6869 }, { "epoch": 0.7424619042472712, "grad_norm": 0.42099326848983765, "learning_rate": 8e-05, "loss": 1.6891, "step": 6870 }, { "epoch": 0.7425699773046579, "grad_norm": 0.41380107402801514, "learning_rate": 8e-05, "loss": 1.6176, "step": 6871 }, { "epoch": 0.7426780503620447, "grad_norm": 0.4033999741077423, "learning_rate": 8e-05, "loss": 1.5282, "step": 6872 }, { "epoch": 0.7427861234194315, "grad_norm": 0.46012306213378906, "learning_rate": 8e-05, "loss": 1.8775, "step": 6873 }, { "epoch": 0.7428941964768183, "grad_norm": 0.437764436006546, "learning_rate": 8e-05, "loss": 1.669, "step": 6874 }, { "epoch": 0.7430022695342051, "grad_norm": 0.4153730869293213, "learning_rate": 8e-05, "loss": 1.597, "step": 6875 }, { "epoch": 0.7431103425915919, "grad_norm": 0.4056847393512726, "learning_rate": 8e-05, "loss": 1.6913, "step": 6876 }, { "epoch": 0.7432184156489787, "grad_norm": 0.4177290201187134, "learning_rate": 8e-05, "loss": 1.6652, "step": 6877 }, { "epoch": 0.7433264887063655, "grad_norm": 0.46593186259269714, "learning_rate": 8e-05, "loss": 1.7707, "step": 6878 }, { "epoch": 0.7434345617637523, "grad_norm": 0.40377798676490784, "learning_rate": 8e-05, "loss": 1.6993, "step": 6879 }, { "epoch": 0.7435426348211391, "grad_norm": 0.428752064704895, "learning_rate": 8e-05, "loss": 1.6891, "step": 6880 }, { "epoch": 0.7436507078785258, "grad_norm": 0.43473613262176514, "learning_rate": 8e-05, "loss": 1.7139, "step": 6881 }, { "epoch": 0.7437587809359126, "grad_norm": 0.4222513437271118, "learning_rate": 8e-05, "loss": 1.781, "step": 6882 }, { "epoch": 0.7438668539932994, "grad_norm": 0.432624876499176, "learning_rate": 8e-05, "loss": 1.7844, "step": 6883 }, { "epoch": 0.7439749270506862, "grad_norm": 0.41381093859672546, "learning_rate": 8e-05, "loss": 1.6534, "step": 6884 }, { "epoch": 0.744083000108073, "grad_norm": 0.4288731515407562, "learning_rate": 8e-05, "loss": 1.659, "step": 6885 }, { "epoch": 0.7441910731654598, "grad_norm": 0.3979169726371765, "learning_rate": 8e-05, "loss": 1.5518, "step": 6886 }, { "epoch": 0.7442991462228467, "grad_norm": 0.43904250860214233, "learning_rate": 8e-05, "loss": 1.7201, "step": 6887 }, { "epoch": 0.7444072192802335, "grad_norm": 0.4336971044540405, "learning_rate": 8e-05, "loss": 1.7128, "step": 6888 }, { "epoch": 0.7445152923376203, "grad_norm": 0.4683052897453308, "learning_rate": 8e-05, "loss": 1.7166, "step": 6889 }, { "epoch": 0.7446233653950071, "grad_norm": 0.4246189594268799, "learning_rate": 8e-05, "loss": 1.7286, "step": 6890 }, { "epoch": 0.7447314384523939, "grad_norm": 0.4129840135574341, "learning_rate": 8e-05, "loss": 1.6326, "step": 6891 }, { "epoch": 0.7448395115097806, "grad_norm": 0.4153437614440918, "learning_rate": 8e-05, "loss": 1.6722, "step": 6892 }, { "epoch": 0.7449475845671674, "grad_norm": 0.44743019342422485, "learning_rate": 8e-05, "loss": 1.6564, "step": 6893 }, { "epoch": 0.7450556576245542, "grad_norm": 0.45590880513191223, "learning_rate": 8e-05, "loss": 1.7148, "step": 6894 }, { "epoch": 0.745163730681941, "grad_norm": 0.46925631165504456, "learning_rate": 8e-05, "loss": 1.8246, "step": 6895 }, { "epoch": 0.7452718037393278, "grad_norm": 0.4582567512989044, "learning_rate": 8e-05, "loss": 1.6609, "step": 6896 }, { "epoch": 0.7453798767967146, "grad_norm": 0.44403165578842163, "learning_rate": 8e-05, "loss": 1.8092, "step": 6897 }, { "epoch": 0.7454879498541014, "grad_norm": 0.45697781443595886, "learning_rate": 8e-05, "loss": 1.7909, "step": 6898 }, { "epoch": 0.7455960229114882, "grad_norm": 0.4248505234718323, "learning_rate": 8e-05, "loss": 1.7286, "step": 6899 }, { "epoch": 0.745704095968875, "grad_norm": 0.3969990909099579, "learning_rate": 8e-05, "loss": 1.6332, "step": 6900 }, { "epoch": 0.7458121690262618, "grad_norm": 0.4832483232021332, "learning_rate": 8e-05, "loss": 1.9524, "step": 6901 }, { "epoch": 0.7459202420836486, "grad_norm": 0.4124661087989807, "learning_rate": 8e-05, "loss": 1.5503, "step": 6902 }, { "epoch": 0.7460283151410353, "grad_norm": 0.42259880900382996, "learning_rate": 8e-05, "loss": 1.6282, "step": 6903 }, { "epoch": 0.7461363881984221, "grad_norm": 0.45760151743888855, "learning_rate": 8e-05, "loss": 1.6214, "step": 6904 }, { "epoch": 0.7462444612558089, "grad_norm": 0.4540753960609436, "learning_rate": 8e-05, "loss": 1.7095, "step": 6905 }, { "epoch": 0.7463525343131957, "grad_norm": 0.4249492585659027, "learning_rate": 8e-05, "loss": 1.6153, "step": 6906 }, { "epoch": 0.7464606073705825, "grad_norm": 0.4273448884487152, "learning_rate": 8e-05, "loss": 1.6828, "step": 6907 }, { "epoch": 0.7465686804279693, "grad_norm": 0.4337843656539917, "learning_rate": 8e-05, "loss": 1.6793, "step": 6908 }, { "epoch": 0.7466767534853561, "grad_norm": 0.4859107434749603, "learning_rate": 8e-05, "loss": 1.6631, "step": 6909 }, { "epoch": 0.7467848265427429, "grad_norm": 0.50737464427948, "learning_rate": 8e-05, "loss": 1.686, "step": 6910 }, { "epoch": 0.7468928996001297, "grad_norm": 0.44387972354888916, "learning_rate": 8e-05, "loss": 1.6013, "step": 6911 }, { "epoch": 0.7470009726575165, "grad_norm": 0.46699219942092896, "learning_rate": 8e-05, "loss": 1.8678, "step": 6912 }, { "epoch": 0.7471090457149032, "grad_norm": 0.4342431426048279, "learning_rate": 8e-05, "loss": 1.608, "step": 6913 }, { "epoch": 0.74721711877229, "grad_norm": 0.3936452567577362, "learning_rate": 8e-05, "loss": 1.6007, "step": 6914 }, { "epoch": 0.7473251918296768, "grad_norm": 0.46810582280158997, "learning_rate": 8e-05, "loss": 1.7143, "step": 6915 }, { "epoch": 0.7474332648870636, "grad_norm": 0.44839245080947876, "learning_rate": 8e-05, "loss": 1.722, "step": 6916 }, { "epoch": 0.7475413379444504, "grad_norm": 0.41502800583839417, "learning_rate": 8e-05, "loss": 1.6606, "step": 6917 }, { "epoch": 0.7476494110018372, "grad_norm": 0.4534761905670166, "learning_rate": 8e-05, "loss": 1.654, "step": 6918 }, { "epoch": 0.747757484059224, "grad_norm": 0.4611088037490845, "learning_rate": 8e-05, "loss": 1.7874, "step": 6919 }, { "epoch": 0.7478655571166108, "grad_norm": 0.4382200539112091, "learning_rate": 8e-05, "loss": 1.7552, "step": 6920 }, { "epoch": 0.7479736301739977, "grad_norm": 0.435476690530777, "learning_rate": 8e-05, "loss": 1.8495, "step": 6921 }, { "epoch": 0.7480817032313845, "grad_norm": 0.49657827615737915, "learning_rate": 8e-05, "loss": 1.4682, "step": 6922 }, { "epoch": 0.7481897762887713, "grad_norm": 0.45124295353889465, "learning_rate": 8e-05, "loss": 1.4956, "step": 6923 }, { "epoch": 0.748297849346158, "grad_norm": 0.40515032410621643, "learning_rate": 8e-05, "loss": 1.5342, "step": 6924 }, { "epoch": 0.7484059224035448, "grad_norm": 0.4344494044780731, "learning_rate": 8e-05, "loss": 1.6243, "step": 6925 }, { "epoch": 0.7485139954609316, "grad_norm": 0.46923092007637024, "learning_rate": 8e-05, "loss": 1.4843, "step": 6926 }, { "epoch": 0.7486220685183184, "grad_norm": 0.47870200872421265, "learning_rate": 8e-05, "loss": 1.8475, "step": 6927 }, { "epoch": 0.7487301415757052, "grad_norm": 0.4126816391944885, "learning_rate": 8e-05, "loss": 1.7331, "step": 6928 }, { "epoch": 0.748838214633092, "grad_norm": 0.49702703952789307, "learning_rate": 8e-05, "loss": 1.7555, "step": 6929 }, { "epoch": 0.7489462876904788, "grad_norm": 0.4526054859161377, "learning_rate": 8e-05, "loss": 1.6791, "step": 6930 }, { "epoch": 0.7490543607478656, "grad_norm": 0.42670080065727234, "learning_rate": 8e-05, "loss": 1.6369, "step": 6931 }, { "epoch": 0.7491624338052524, "grad_norm": 0.4506491422653198, "learning_rate": 8e-05, "loss": 1.7498, "step": 6932 }, { "epoch": 0.7492705068626392, "grad_norm": 0.4451931118965149, "learning_rate": 8e-05, "loss": 1.8031, "step": 6933 }, { "epoch": 0.749378579920026, "grad_norm": 0.5178097486495972, "learning_rate": 8e-05, "loss": 1.7891, "step": 6934 }, { "epoch": 0.7494866529774127, "grad_norm": 0.4346865117549896, "learning_rate": 8e-05, "loss": 1.821, "step": 6935 }, { "epoch": 0.7495947260347995, "grad_norm": 0.4289848804473877, "learning_rate": 8e-05, "loss": 1.7075, "step": 6936 }, { "epoch": 0.7497027990921863, "grad_norm": 0.40796443819999695, "learning_rate": 8e-05, "loss": 1.5446, "step": 6937 }, { "epoch": 0.7498108721495731, "grad_norm": 0.4874688982963562, "learning_rate": 8e-05, "loss": 1.5241, "step": 6938 }, { "epoch": 0.7499189452069599, "grad_norm": 0.43724748492240906, "learning_rate": 8e-05, "loss": 1.7403, "step": 6939 }, { "epoch": 0.7500270182643467, "grad_norm": 0.41095924377441406, "learning_rate": 8e-05, "loss": 1.6574, "step": 6940 }, { "epoch": 0.7501350913217335, "grad_norm": 0.4294879734516144, "learning_rate": 8e-05, "loss": 1.727, "step": 6941 }, { "epoch": 0.7502431643791203, "grad_norm": 0.4114168882369995, "learning_rate": 8e-05, "loss": 1.7566, "step": 6942 }, { "epoch": 0.7503512374365071, "grad_norm": 0.43039748072624207, "learning_rate": 8e-05, "loss": 1.7488, "step": 6943 }, { "epoch": 0.7504593104938939, "grad_norm": 0.44461873173713684, "learning_rate": 8e-05, "loss": 1.8038, "step": 6944 }, { "epoch": 0.7505673835512807, "grad_norm": 0.44380319118499756, "learning_rate": 8e-05, "loss": 1.7841, "step": 6945 }, { "epoch": 0.7506754566086674, "grad_norm": 0.48491886258125305, "learning_rate": 8e-05, "loss": 1.9473, "step": 6946 }, { "epoch": 0.7507835296660542, "grad_norm": 0.41113030910491943, "learning_rate": 8e-05, "loss": 1.5105, "step": 6947 }, { "epoch": 0.750891602723441, "grad_norm": 0.4194891154766083, "learning_rate": 8e-05, "loss": 1.6767, "step": 6948 }, { "epoch": 0.7509996757808278, "grad_norm": 0.48622390627861023, "learning_rate": 8e-05, "loss": 1.8822, "step": 6949 }, { "epoch": 0.7511077488382146, "grad_norm": 0.40224093198776245, "learning_rate": 8e-05, "loss": 1.4801, "step": 6950 }, { "epoch": 0.7512158218956014, "grad_norm": 0.46113499999046326, "learning_rate": 8e-05, "loss": 1.6912, "step": 6951 }, { "epoch": 0.7513238949529882, "grad_norm": 0.4231925308704376, "learning_rate": 8e-05, "loss": 1.6269, "step": 6952 }, { "epoch": 0.751431968010375, "grad_norm": 0.39638176560401917, "learning_rate": 8e-05, "loss": 1.5067, "step": 6953 }, { "epoch": 0.7515400410677618, "grad_norm": 0.4263779819011688, "learning_rate": 8e-05, "loss": 1.688, "step": 6954 }, { "epoch": 0.7516481141251486, "grad_norm": 0.4364466667175293, "learning_rate": 8e-05, "loss": 1.638, "step": 6955 }, { "epoch": 0.7517561871825353, "grad_norm": 0.43693697452545166, "learning_rate": 8e-05, "loss": 1.6838, "step": 6956 }, { "epoch": 0.7518642602399221, "grad_norm": 0.46510449051856995, "learning_rate": 8e-05, "loss": 1.8072, "step": 6957 }, { "epoch": 0.751972333297309, "grad_norm": 0.4119250476360321, "learning_rate": 8e-05, "loss": 1.7281, "step": 6958 }, { "epoch": 0.7520804063546958, "grad_norm": 0.4692193269729614, "learning_rate": 8e-05, "loss": 1.8627, "step": 6959 }, { "epoch": 0.7521884794120826, "grad_norm": 0.430829256772995, "learning_rate": 8e-05, "loss": 1.746, "step": 6960 }, { "epoch": 0.7522965524694694, "grad_norm": 0.4096299409866333, "learning_rate": 8e-05, "loss": 1.6077, "step": 6961 }, { "epoch": 0.7524046255268562, "grad_norm": 0.44863253831863403, "learning_rate": 8e-05, "loss": 1.7109, "step": 6962 }, { "epoch": 0.752512698584243, "grad_norm": 0.4820490777492523, "learning_rate": 8e-05, "loss": 1.6339, "step": 6963 }, { "epoch": 0.7526207716416298, "grad_norm": 0.40907683968544006, "learning_rate": 8e-05, "loss": 1.6904, "step": 6964 }, { "epoch": 0.7527288446990166, "grad_norm": 0.419467031955719, "learning_rate": 8e-05, "loss": 1.7282, "step": 6965 }, { "epoch": 0.7528369177564034, "grad_norm": 0.4190618097782135, "learning_rate": 8e-05, "loss": 1.7077, "step": 6966 }, { "epoch": 0.7529449908137901, "grad_norm": 0.5079578161239624, "learning_rate": 8e-05, "loss": 1.8836, "step": 6967 }, { "epoch": 0.7530530638711769, "grad_norm": 0.454111248254776, "learning_rate": 8e-05, "loss": 1.5833, "step": 6968 }, { "epoch": 0.7531611369285637, "grad_norm": 0.4300159811973572, "learning_rate": 8e-05, "loss": 1.7374, "step": 6969 }, { "epoch": 0.7532692099859505, "grad_norm": 0.4583958387374878, "learning_rate": 8e-05, "loss": 1.5809, "step": 6970 }, { "epoch": 0.7533772830433373, "grad_norm": 0.4049728214740753, "learning_rate": 8e-05, "loss": 1.5993, "step": 6971 }, { "epoch": 0.7534853561007241, "grad_norm": 0.432932585477829, "learning_rate": 8e-05, "loss": 1.7405, "step": 6972 }, { "epoch": 0.7535934291581109, "grad_norm": 0.478544682264328, "learning_rate": 8e-05, "loss": 1.818, "step": 6973 }, { "epoch": 0.7537015022154977, "grad_norm": 0.40245670080184937, "learning_rate": 8e-05, "loss": 1.6396, "step": 6974 }, { "epoch": 0.7538095752728845, "grad_norm": 0.42570704221725464, "learning_rate": 8e-05, "loss": 1.6087, "step": 6975 }, { "epoch": 0.7539176483302713, "grad_norm": 0.4944288432598114, "learning_rate": 8e-05, "loss": 1.7496, "step": 6976 }, { "epoch": 0.7540257213876581, "grad_norm": 0.409181147813797, "learning_rate": 8e-05, "loss": 1.6505, "step": 6977 }, { "epoch": 0.7541337944450448, "grad_norm": 0.42837241291999817, "learning_rate": 8e-05, "loss": 1.638, "step": 6978 }, { "epoch": 0.7542418675024316, "grad_norm": 0.42104044556617737, "learning_rate": 8e-05, "loss": 1.6915, "step": 6979 }, { "epoch": 0.7543499405598184, "grad_norm": 0.46025991439819336, "learning_rate": 8e-05, "loss": 1.9131, "step": 6980 }, { "epoch": 0.7544580136172052, "grad_norm": 0.5014327168464661, "learning_rate": 8e-05, "loss": 1.928, "step": 6981 }, { "epoch": 0.754566086674592, "grad_norm": 0.4150974154472351, "learning_rate": 8e-05, "loss": 1.688, "step": 6982 }, { "epoch": 0.7546741597319788, "grad_norm": 0.4447047710418701, "learning_rate": 8e-05, "loss": 1.6069, "step": 6983 }, { "epoch": 0.7547822327893656, "grad_norm": 0.43278932571411133, "learning_rate": 8e-05, "loss": 1.6323, "step": 6984 }, { "epoch": 0.7548903058467524, "grad_norm": 0.42932137846946716, "learning_rate": 8e-05, "loss": 1.63, "step": 6985 }, { "epoch": 0.7549983789041392, "grad_norm": 0.3868107795715332, "learning_rate": 8e-05, "loss": 1.5807, "step": 6986 }, { "epoch": 0.755106451961526, "grad_norm": 0.42482003569602966, "learning_rate": 8e-05, "loss": 1.5946, "step": 6987 }, { "epoch": 0.7552145250189127, "grad_norm": 0.4004952907562256, "learning_rate": 8e-05, "loss": 1.6803, "step": 6988 }, { "epoch": 0.7553225980762995, "grad_norm": 0.4974971413612366, "learning_rate": 8e-05, "loss": 1.8414, "step": 6989 }, { "epoch": 0.7554306711336863, "grad_norm": 0.4780620336532593, "learning_rate": 8e-05, "loss": 1.7315, "step": 6990 }, { "epoch": 0.7555387441910731, "grad_norm": 0.4600673019886017, "learning_rate": 8e-05, "loss": 1.6734, "step": 6991 }, { "epoch": 0.75564681724846, "grad_norm": 0.429053395986557, "learning_rate": 8e-05, "loss": 1.4742, "step": 6992 }, { "epoch": 0.7557548903058467, "grad_norm": 0.46207356452941895, "learning_rate": 8e-05, "loss": 1.7828, "step": 6993 }, { "epoch": 0.7558629633632336, "grad_norm": 0.4188014566898346, "learning_rate": 8e-05, "loss": 1.7188, "step": 6994 }, { "epoch": 0.7559710364206204, "grad_norm": 0.49204733967781067, "learning_rate": 8e-05, "loss": 1.8898, "step": 6995 }, { "epoch": 0.7560791094780072, "grad_norm": 0.46167340874671936, "learning_rate": 8e-05, "loss": 1.7276, "step": 6996 }, { "epoch": 0.756187182535394, "grad_norm": 0.44232794642448425, "learning_rate": 8e-05, "loss": 1.771, "step": 6997 }, { "epoch": 0.7562952555927808, "grad_norm": 0.43187251687049866, "learning_rate": 8e-05, "loss": 1.6852, "step": 6998 }, { "epoch": 0.7564033286501675, "grad_norm": 0.4343647360801697, "learning_rate": 8e-05, "loss": 1.7133, "step": 6999 }, { "epoch": 0.7565114017075543, "grad_norm": 0.4346638023853302, "learning_rate": 8e-05, "loss": 1.7689, "step": 7000 }, { "epoch": 0.7566194747649411, "grad_norm": 0.45481932163238525, "learning_rate": 8e-05, "loss": 1.829, "step": 7001 }, { "epoch": 0.7567275478223279, "grad_norm": 0.44893839955329895, "learning_rate": 8e-05, "loss": 1.8285, "step": 7002 }, { "epoch": 0.7568356208797147, "grad_norm": 0.4102190136909485, "learning_rate": 8e-05, "loss": 1.5636, "step": 7003 }, { "epoch": 0.7569436939371015, "grad_norm": 0.3869622051715851, "learning_rate": 8e-05, "loss": 1.5852, "step": 7004 }, { "epoch": 0.7570517669944883, "grad_norm": 0.38854217529296875, "learning_rate": 8e-05, "loss": 1.5664, "step": 7005 }, { "epoch": 0.7571598400518751, "grad_norm": 0.4142291843891144, "learning_rate": 8e-05, "loss": 1.7819, "step": 7006 }, { "epoch": 0.7572679131092619, "grad_norm": 0.41838034987449646, "learning_rate": 8e-05, "loss": 1.6076, "step": 7007 }, { "epoch": 0.7573759861666487, "grad_norm": 0.43625709414482117, "learning_rate": 8e-05, "loss": 1.7442, "step": 7008 }, { "epoch": 0.7574840592240355, "grad_norm": 0.47639626264572144, "learning_rate": 8e-05, "loss": 1.8295, "step": 7009 }, { "epoch": 0.7575921322814222, "grad_norm": 0.4679754972457886, "learning_rate": 8e-05, "loss": 1.599, "step": 7010 }, { "epoch": 0.757700205338809, "grad_norm": 0.43693456053733826, "learning_rate": 8e-05, "loss": 1.6747, "step": 7011 }, { "epoch": 0.7578082783961958, "grad_norm": 0.409050315618515, "learning_rate": 8e-05, "loss": 1.6418, "step": 7012 }, { "epoch": 0.7579163514535826, "grad_norm": 0.4482717216014862, "learning_rate": 8e-05, "loss": 1.5598, "step": 7013 }, { "epoch": 0.7580244245109694, "grad_norm": 0.4133572280406952, "learning_rate": 8e-05, "loss": 1.7148, "step": 7014 }, { "epoch": 0.7581324975683562, "grad_norm": 0.43680188059806824, "learning_rate": 8e-05, "loss": 1.7487, "step": 7015 }, { "epoch": 0.758240570625743, "grad_norm": 0.46906426548957825, "learning_rate": 8e-05, "loss": 1.6798, "step": 7016 }, { "epoch": 0.7583486436831298, "grad_norm": 0.4509083926677704, "learning_rate": 8e-05, "loss": 1.8532, "step": 7017 }, { "epoch": 0.7584567167405166, "grad_norm": 0.4290767014026642, "learning_rate": 8e-05, "loss": 1.6345, "step": 7018 }, { "epoch": 0.7585647897979034, "grad_norm": 0.42889168858528137, "learning_rate": 8e-05, "loss": 1.5692, "step": 7019 }, { "epoch": 0.7586728628552902, "grad_norm": 0.4440365731716156, "learning_rate": 8e-05, "loss": 1.863, "step": 7020 }, { "epoch": 0.7587809359126769, "grad_norm": 0.4699527323246002, "learning_rate": 8e-05, "loss": 1.777, "step": 7021 }, { "epoch": 0.7588890089700637, "grad_norm": 0.4183804392814636, "learning_rate": 8e-05, "loss": 1.6569, "step": 7022 }, { "epoch": 0.7589970820274505, "grad_norm": 0.5121611952781677, "learning_rate": 8e-05, "loss": 1.7896, "step": 7023 }, { "epoch": 0.7591051550848373, "grad_norm": 0.448169469833374, "learning_rate": 8e-05, "loss": 1.5239, "step": 7024 }, { "epoch": 0.7592132281422241, "grad_norm": 0.4759168028831482, "learning_rate": 8e-05, "loss": 1.6236, "step": 7025 }, { "epoch": 0.7593213011996109, "grad_norm": 0.4584440290927887, "learning_rate": 8e-05, "loss": 1.8283, "step": 7026 }, { "epoch": 0.7594293742569977, "grad_norm": 0.43264544010162354, "learning_rate": 8e-05, "loss": 1.6929, "step": 7027 }, { "epoch": 0.7595374473143846, "grad_norm": 0.43830251693725586, "learning_rate": 8e-05, "loss": 1.6293, "step": 7028 }, { "epoch": 0.7596455203717714, "grad_norm": 0.4066115915775299, "learning_rate": 8e-05, "loss": 1.6683, "step": 7029 }, { "epoch": 0.7597535934291582, "grad_norm": 0.4621807932853699, "learning_rate": 8e-05, "loss": 1.756, "step": 7030 }, { "epoch": 0.7598616664865449, "grad_norm": 0.4291408658027649, "learning_rate": 8e-05, "loss": 1.7317, "step": 7031 }, { "epoch": 0.7599697395439317, "grad_norm": 0.3847090005874634, "learning_rate": 8e-05, "loss": 1.5038, "step": 7032 }, { "epoch": 0.7600778126013185, "grad_norm": 0.406271368265152, "learning_rate": 8e-05, "loss": 1.6543, "step": 7033 }, { "epoch": 0.7601858856587053, "grad_norm": 0.446780800819397, "learning_rate": 8e-05, "loss": 1.712, "step": 7034 }, { "epoch": 0.7602939587160921, "grad_norm": 0.43409019708633423, "learning_rate": 8e-05, "loss": 1.7779, "step": 7035 }, { "epoch": 0.7604020317734789, "grad_norm": 0.4331832826137543, "learning_rate": 8e-05, "loss": 1.7209, "step": 7036 }, { "epoch": 0.7605101048308657, "grad_norm": 0.44164884090423584, "learning_rate": 8e-05, "loss": 1.8455, "step": 7037 }, { "epoch": 0.7606181778882525, "grad_norm": 0.4007634222507477, "learning_rate": 8e-05, "loss": 1.3901, "step": 7038 }, { "epoch": 0.7607262509456393, "grad_norm": 0.42758703231811523, "learning_rate": 8e-05, "loss": 1.7075, "step": 7039 }, { "epoch": 0.7608343240030261, "grad_norm": 0.4046802520751953, "learning_rate": 8e-05, "loss": 1.6777, "step": 7040 }, { "epoch": 0.7609423970604129, "grad_norm": 0.42277631163597107, "learning_rate": 8e-05, "loss": 1.753, "step": 7041 }, { "epoch": 0.7610504701177996, "grad_norm": 0.4140164256095886, "learning_rate": 8e-05, "loss": 1.7081, "step": 7042 }, { "epoch": 0.7611585431751864, "grad_norm": 0.4381808936595917, "learning_rate": 8e-05, "loss": 1.6731, "step": 7043 }, { "epoch": 0.7612666162325732, "grad_norm": 0.4466732144355774, "learning_rate": 8e-05, "loss": 1.7439, "step": 7044 }, { "epoch": 0.76137468928996, "grad_norm": 0.39488646388053894, "learning_rate": 8e-05, "loss": 1.4394, "step": 7045 }, { "epoch": 0.7614827623473468, "grad_norm": 0.43336060643196106, "learning_rate": 8e-05, "loss": 1.7969, "step": 7046 }, { "epoch": 0.7615908354047336, "grad_norm": 0.4294263422489166, "learning_rate": 8e-05, "loss": 1.6836, "step": 7047 }, { "epoch": 0.7616989084621204, "grad_norm": 0.456287145614624, "learning_rate": 8e-05, "loss": 1.6943, "step": 7048 }, { "epoch": 0.7618069815195072, "grad_norm": 0.5845528841018677, "learning_rate": 8e-05, "loss": 1.8177, "step": 7049 }, { "epoch": 0.761915054576894, "grad_norm": 0.43678781390190125, "learning_rate": 8e-05, "loss": 1.4482, "step": 7050 }, { "epoch": 0.7620231276342808, "grad_norm": 0.44243717193603516, "learning_rate": 8e-05, "loss": 1.7925, "step": 7051 }, { "epoch": 0.7621312006916676, "grad_norm": 0.5242395997047424, "learning_rate": 8e-05, "loss": 1.8359, "step": 7052 }, { "epoch": 0.7622392737490543, "grad_norm": 0.4326714277267456, "learning_rate": 8e-05, "loss": 1.8437, "step": 7053 }, { "epoch": 0.7623473468064411, "grad_norm": 0.4641999304294586, "learning_rate": 8e-05, "loss": 1.735, "step": 7054 }, { "epoch": 0.7624554198638279, "grad_norm": 0.460612028837204, "learning_rate": 8e-05, "loss": 1.7674, "step": 7055 }, { "epoch": 0.7625634929212147, "grad_norm": 0.40984466671943665, "learning_rate": 8e-05, "loss": 1.6167, "step": 7056 }, { "epoch": 0.7626715659786015, "grad_norm": 0.45699042081832886, "learning_rate": 8e-05, "loss": 1.7454, "step": 7057 }, { "epoch": 0.7627796390359883, "grad_norm": 0.4250757396221161, "learning_rate": 8e-05, "loss": 1.6949, "step": 7058 }, { "epoch": 0.7628877120933751, "grad_norm": 0.5017492771148682, "learning_rate": 8e-05, "loss": 1.7729, "step": 7059 }, { "epoch": 0.7629957851507619, "grad_norm": 0.4162467122077942, "learning_rate": 8e-05, "loss": 1.6799, "step": 7060 }, { "epoch": 0.7631038582081487, "grad_norm": 0.43121668696403503, "learning_rate": 8e-05, "loss": 1.6166, "step": 7061 }, { "epoch": 0.7632119312655355, "grad_norm": 0.44202330708503723, "learning_rate": 8e-05, "loss": 1.7794, "step": 7062 }, { "epoch": 0.7633200043229222, "grad_norm": 0.42845380306243896, "learning_rate": 8e-05, "loss": 1.7096, "step": 7063 }, { "epoch": 0.763428077380309, "grad_norm": 0.440114825963974, "learning_rate": 8e-05, "loss": 1.8145, "step": 7064 }, { "epoch": 0.7635361504376958, "grad_norm": 0.4988890290260315, "learning_rate": 8e-05, "loss": 1.9724, "step": 7065 }, { "epoch": 0.7636442234950827, "grad_norm": 0.412405401468277, "learning_rate": 8e-05, "loss": 1.4782, "step": 7066 }, { "epoch": 0.7637522965524695, "grad_norm": 0.44073686003685, "learning_rate": 8e-05, "loss": 1.5308, "step": 7067 }, { "epoch": 0.7638603696098563, "grad_norm": 0.45187491178512573, "learning_rate": 8e-05, "loss": 1.6493, "step": 7068 }, { "epoch": 0.7639684426672431, "grad_norm": 0.4300079345703125, "learning_rate": 8e-05, "loss": 1.718, "step": 7069 }, { "epoch": 0.7640765157246299, "grad_norm": 0.47499746084213257, "learning_rate": 8e-05, "loss": 1.6498, "step": 7070 }, { "epoch": 0.7641845887820167, "grad_norm": 0.4173726439476013, "learning_rate": 8e-05, "loss": 1.4708, "step": 7071 }, { "epoch": 0.7642926618394035, "grad_norm": 0.455045610666275, "learning_rate": 8e-05, "loss": 1.6252, "step": 7072 }, { "epoch": 0.7644007348967903, "grad_norm": 0.4672051668167114, "learning_rate": 8e-05, "loss": 1.845, "step": 7073 }, { "epoch": 0.764508807954177, "grad_norm": 0.46072250604629517, "learning_rate": 8e-05, "loss": 1.8126, "step": 7074 }, { "epoch": 0.7646168810115638, "grad_norm": 0.4124811589717865, "learning_rate": 8e-05, "loss": 1.4454, "step": 7075 }, { "epoch": 0.7647249540689506, "grad_norm": 0.4791329801082611, "learning_rate": 8e-05, "loss": 1.7578, "step": 7076 }, { "epoch": 0.7648330271263374, "grad_norm": 0.46811842918395996, "learning_rate": 8e-05, "loss": 1.693, "step": 7077 }, { "epoch": 0.7649411001837242, "grad_norm": 0.4314934015274048, "learning_rate": 8e-05, "loss": 1.7399, "step": 7078 }, { "epoch": 0.765049173241111, "grad_norm": 0.4523933231830597, "learning_rate": 8e-05, "loss": 1.8963, "step": 7079 }, { "epoch": 0.7651572462984978, "grad_norm": 0.45262351632118225, "learning_rate": 8e-05, "loss": 1.4046, "step": 7080 }, { "epoch": 0.7652653193558846, "grad_norm": 0.4707155227661133, "learning_rate": 8e-05, "loss": 1.7782, "step": 7081 }, { "epoch": 0.7653733924132714, "grad_norm": 0.43137985467910767, "learning_rate": 8e-05, "loss": 1.5318, "step": 7082 }, { "epoch": 0.7654814654706582, "grad_norm": 0.4180015027523041, "learning_rate": 8e-05, "loss": 1.5578, "step": 7083 }, { "epoch": 0.765589538528045, "grad_norm": 0.4436553120613098, "learning_rate": 8e-05, "loss": 1.7631, "step": 7084 }, { "epoch": 0.7656976115854317, "grad_norm": 0.4088929295539856, "learning_rate": 8e-05, "loss": 1.5436, "step": 7085 }, { "epoch": 0.7658056846428185, "grad_norm": 0.4112737476825714, "learning_rate": 8e-05, "loss": 1.7043, "step": 7086 }, { "epoch": 0.7659137577002053, "grad_norm": 0.4892650544643402, "learning_rate": 8e-05, "loss": 1.8146, "step": 7087 }, { "epoch": 0.7660218307575921, "grad_norm": 0.43112555146217346, "learning_rate": 8e-05, "loss": 1.5302, "step": 7088 }, { "epoch": 0.7661299038149789, "grad_norm": 0.4553471803665161, "learning_rate": 8e-05, "loss": 1.8543, "step": 7089 }, { "epoch": 0.7662379768723657, "grad_norm": 0.42358672618865967, "learning_rate": 8e-05, "loss": 1.6661, "step": 7090 }, { "epoch": 0.7663460499297525, "grad_norm": 0.4471980035305023, "learning_rate": 8e-05, "loss": 1.6927, "step": 7091 }, { "epoch": 0.7664541229871393, "grad_norm": 0.42600613832473755, "learning_rate": 8e-05, "loss": 1.8157, "step": 7092 }, { "epoch": 0.7665621960445261, "grad_norm": 0.4266655743122101, "learning_rate": 8e-05, "loss": 1.584, "step": 7093 }, { "epoch": 0.7666702691019129, "grad_norm": 0.4235292375087738, "learning_rate": 8e-05, "loss": 1.6426, "step": 7094 }, { "epoch": 0.7667783421592997, "grad_norm": 0.43275579810142517, "learning_rate": 8e-05, "loss": 1.6675, "step": 7095 }, { "epoch": 0.7668864152166864, "grad_norm": 0.44073861837387085, "learning_rate": 8e-05, "loss": 1.6279, "step": 7096 }, { "epoch": 0.7669944882740732, "grad_norm": 0.4491444230079651, "learning_rate": 8e-05, "loss": 1.7266, "step": 7097 }, { "epoch": 0.76710256133146, "grad_norm": 0.4073023200035095, "learning_rate": 8e-05, "loss": 1.4224, "step": 7098 }, { "epoch": 0.7672106343888468, "grad_norm": 0.4202984869480133, "learning_rate": 8e-05, "loss": 1.654, "step": 7099 }, { "epoch": 0.7673187074462336, "grad_norm": 0.4011022448539734, "learning_rate": 8e-05, "loss": 1.5762, "step": 7100 }, { "epoch": 0.7674267805036205, "grad_norm": 0.4559363126754761, "learning_rate": 8e-05, "loss": 1.7012, "step": 7101 }, { "epoch": 0.7675348535610073, "grad_norm": 0.45984822511672974, "learning_rate": 8e-05, "loss": 1.7231, "step": 7102 }, { "epoch": 0.7676429266183941, "grad_norm": 0.44693082571029663, "learning_rate": 8e-05, "loss": 1.632, "step": 7103 }, { "epoch": 0.7677509996757809, "grad_norm": 0.41280993819236755, "learning_rate": 8e-05, "loss": 1.6139, "step": 7104 }, { "epoch": 0.7678590727331677, "grad_norm": 0.4881168007850647, "learning_rate": 8e-05, "loss": 1.7223, "step": 7105 }, { "epoch": 0.7679671457905544, "grad_norm": 0.4466334283351898, "learning_rate": 8e-05, "loss": 1.8017, "step": 7106 }, { "epoch": 0.7680752188479412, "grad_norm": 0.4113959074020386, "learning_rate": 8e-05, "loss": 1.6376, "step": 7107 }, { "epoch": 0.768183291905328, "grad_norm": 0.43910491466522217, "learning_rate": 8e-05, "loss": 1.6167, "step": 7108 }, { "epoch": 0.7682913649627148, "grad_norm": 0.4416026771068573, "learning_rate": 8e-05, "loss": 1.8436, "step": 7109 }, { "epoch": 0.7683994380201016, "grad_norm": 0.41756874322891235, "learning_rate": 8e-05, "loss": 1.5732, "step": 7110 }, { "epoch": 0.7685075110774884, "grad_norm": 0.39522144198417664, "learning_rate": 8e-05, "loss": 1.6533, "step": 7111 }, { "epoch": 0.7686155841348752, "grad_norm": 0.4340183436870575, "learning_rate": 8e-05, "loss": 1.7916, "step": 7112 }, { "epoch": 0.768723657192262, "grad_norm": 0.40632444620132446, "learning_rate": 8e-05, "loss": 1.7382, "step": 7113 }, { "epoch": 0.7688317302496488, "grad_norm": 0.44639408588409424, "learning_rate": 8e-05, "loss": 1.6635, "step": 7114 }, { "epoch": 0.7689398033070356, "grad_norm": 0.4345468282699585, "learning_rate": 8e-05, "loss": 1.7209, "step": 7115 }, { "epoch": 0.7690478763644224, "grad_norm": 0.3879147469997406, "learning_rate": 8e-05, "loss": 1.4814, "step": 7116 }, { "epoch": 0.7691559494218091, "grad_norm": 0.42649298906326294, "learning_rate": 8e-05, "loss": 1.7514, "step": 7117 }, { "epoch": 0.7692640224791959, "grad_norm": 0.47357064485549927, "learning_rate": 8e-05, "loss": 1.6357, "step": 7118 }, { "epoch": 0.7693720955365827, "grad_norm": 0.45714572072029114, "learning_rate": 8e-05, "loss": 1.8201, "step": 7119 }, { "epoch": 0.7694801685939695, "grad_norm": 0.44432201981544495, "learning_rate": 8e-05, "loss": 1.6965, "step": 7120 }, { "epoch": 0.7695882416513563, "grad_norm": 0.4585128426551819, "learning_rate": 8e-05, "loss": 1.8024, "step": 7121 }, { "epoch": 0.7696963147087431, "grad_norm": 0.45124760270118713, "learning_rate": 8e-05, "loss": 1.6131, "step": 7122 }, { "epoch": 0.7698043877661299, "grad_norm": 0.4251530170440674, "learning_rate": 8e-05, "loss": 1.6689, "step": 7123 }, { "epoch": 0.7699124608235167, "grad_norm": 0.45423558354377747, "learning_rate": 8e-05, "loss": 1.6862, "step": 7124 }, { "epoch": 0.7700205338809035, "grad_norm": 0.5161629915237427, "learning_rate": 8e-05, "loss": 1.9285, "step": 7125 }, { "epoch": 0.7701286069382903, "grad_norm": 0.4345954656600952, "learning_rate": 8e-05, "loss": 1.6134, "step": 7126 }, { "epoch": 0.7702366799956771, "grad_norm": 0.5196511149406433, "learning_rate": 8e-05, "loss": 1.549, "step": 7127 }, { "epoch": 0.7703447530530638, "grad_norm": 0.4447364807128906, "learning_rate": 8e-05, "loss": 1.8059, "step": 7128 }, { "epoch": 0.7704528261104506, "grad_norm": 0.41200020909309387, "learning_rate": 8e-05, "loss": 1.533, "step": 7129 }, { "epoch": 0.7705608991678374, "grad_norm": 0.42673709988594055, "learning_rate": 8e-05, "loss": 1.633, "step": 7130 }, { "epoch": 0.7706689722252242, "grad_norm": 0.41385480761528015, "learning_rate": 8e-05, "loss": 1.5353, "step": 7131 }, { "epoch": 0.770777045282611, "grad_norm": 0.5009647607803345, "learning_rate": 8e-05, "loss": 1.7813, "step": 7132 }, { "epoch": 0.7708851183399978, "grad_norm": 0.4253973662853241, "learning_rate": 8e-05, "loss": 1.6607, "step": 7133 }, { "epoch": 0.7709931913973846, "grad_norm": 0.4678027629852295, "learning_rate": 8e-05, "loss": 1.6459, "step": 7134 }, { "epoch": 0.7711012644547715, "grad_norm": 0.4352409839630127, "learning_rate": 8e-05, "loss": 1.7509, "step": 7135 }, { "epoch": 0.7712093375121583, "grad_norm": 0.4289982318878174, "learning_rate": 8e-05, "loss": 1.7228, "step": 7136 }, { "epoch": 0.7713174105695451, "grad_norm": 0.422301709651947, "learning_rate": 8e-05, "loss": 1.6183, "step": 7137 }, { "epoch": 0.7714254836269318, "grad_norm": 0.4007168710231781, "learning_rate": 8e-05, "loss": 1.4971, "step": 7138 }, { "epoch": 0.7715335566843186, "grad_norm": 0.38832786679267883, "learning_rate": 8e-05, "loss": 1.3925, "step": 7139 }, { "epoch": 0.7716416297417054, "grad_norm": 0.42680227756500244, "learning_rate": 8e-05, "loss": 1.7334, "step": 7140 }, { "epoch": 0.7717497027990922, "grad_norm": 0.45529404282569885, "learning_rate": 8e-05, "loss": 1.8276, "step": 7141 }, { "epoch": 0.771857775856479, "grad_norm": 0.49499884247779846, "learning_rate": 8e-05, "loss": 1.8472, "step": 7142 }, { "epoch": 0.7719658489138658, "grad_norm": 0.4583873450756073, "learning_rate": 8e-05, "loss": 1.6613, "step": 7143 }, { "epoch": 0.7720739219712526, "grad_norm": 0.4202198386192322, "learning_rate": 8e-05, "loss": 1.4969, "step": 7144 }, { "epoch": 0.7721819950286394, "grad_norm": 0.46239161491394043, "learning_rate": 8e-05, "loss": 1.7195, "step": 7145 }, { "epoch": 0.7722900680860262, "grad_norm": 0.4418841302394867, "learning_rate": 8e-05, "loss": 1.4145, "step": 7146 }, { "epoch": 0.772398141143413, "grad_norm": 0.4434622824192047, "learning_rate": 8e-05, "loss": 1.649, "step": 7147 }, { "epoch": 0.7725062142007998, "grad_norm": 0.4443179666996002, "learning_rate": 8e-05, "loss": 1.7806, "step": 7148 }, { "epoch": 0.7726142872581865, "grad_norm": 0.41300973296165466, "learning_rate": 8e-05, "loss": 1.5255, "step": 7149 }, { "epoch": 0.7727223603155733, "grad_norm": 0.43779581785202026, "learning_rate": 8e-05, "loss": 1.6508, "step": 7150 }, { "epoch": 0.7728304333729601, "grad_norm": 0.4741148054599762, "learning_rate": 8e-05, "loss": 1.7194, "step": 7151 }, { "epoch": 0.7729385064303469, "grad_norm": 0.45650961995124817, "learning_rate": 8e-05, "loss": 1.6483, "step": 7152 }, { "epoch": 0.7730465794877337, "grad_norm": 0.45482486486434937, "learning_rate": 8e-05, "loss": 1.7667, "step": 7153 }, { "epoch": 0.7731546525451205, "grad_norm": 0.4234373867511749, "learning_rate": 8e-05, "loss": 1.7277, "step": 7154 }, { "epoch": 0.7732627256025073, "grad_norm": 0.525168776512146, "learning_rate": 8e-05, "loss": 1.6452, "step": 7155 }, { "epoch": 0.7733707986598941, "grad_norm": 0.44177985191345215, "learning_rate": 8e-05, "loss": 1.6439, "step": 7156 }, { "epoch": 0.7734788717172809, "grad_norm": 0.46131396293640137, "learning_rate": 8e-05, "loss": 1.8644, "step": 7157 }, { "epoch": 0.7735869447746677, "grad_norm": 0.4618714153766632, "learning_rate": 8e-05, "loss": 1.6879, "step": 7158 }, { "epoch": 0.7736950178320545, "grad_norm": 0.44554200768470764, "learning_rate": 8e-05, "loss": 1.7939, "step": 7159 }, { "epoch": 0.7738030908894412, "grad_norm": 0.41176339983940125, "learning_rate": 8e-05, "loss": 1.7426, "step": 7160 }, { "epoch": 0.773911163946828, "grad_norm": 0.4686107933521271, "learning_rate": 8e-05, "loss": 1.8527, "step": 7161 }, { "epoch": 0.7740192370042148, "grad_norm": 0.4289707839488983, "learning_rate": 8e-05, "loss": 1.6852, "step": 7162 }, { "epoch": 0.7741273100616016, "grad_norm": 0.4484335482120514, "learning_rate": 8e-05, "loss": 1.6747, "step": 7163 }, { "epoch": 0.7742353831189884, "grad_norm": 0.4198009967803955, "learning_rate": 8e-05, "loss": 1.6904, "step": 7164 }, { "epoch": 0.7743434561763752, "grad_norm": 0.4186418652534485, "learning_rate": 8e-05, "loss": 1.5759, "step": 7165 }, { "epoch": 0.774451529233762, "grad_norm": 0.42738088965415955, "learning_rate": 8e-05, "loss": 1.5056, "step": 7166 }, { "epoch": 0.7745596022911488, "grad_norm": 0.44802337884902954, "learning_rate": 8e-05, "loss": 1.7363, "step": 7167 }, { "epoch": 0.7746676753485356, "grad_norm": 0.4034801721572876, "learning_rate": 8e-05, "loss": 1.5648, "step": 7168 }, { "epoch": 0.7747757484059224, "grad_norm": 0.5260096788406372, "learning_rate": 8e-05, "loss": 1.8863, "step": 7169 }, { "epoch": 0.7748838214633093, "grad_norm": 0.42227575182914734, "learning_rate": 8e-05, "loss": 1.6162, "step": 7170 }, { "epoch": 0.774991894520696, "grad_norm": 0.41266778111457825, "learning_rate": 8e-05, "loss": 1.545, "step": 7171 }, { "epoch": 0.7750999675780827, "grad_norm": 0.4749162495136261, "learning_rate": 8e-05, "loss": 1.7594, "step": 7172 }, { "epoch": 0.7752080406354696, "grad_norm": 0.3989289700984955, "learning_rate": 8e-05, "loss": 1.4368, "step": 7173 }, { "epoch": 0.7753161136928564, "grad_norm": 0.4655546247959137, "learning_rate": 8e-05, "loss": 1.6049, "step": 7174 }, { "epoch": 0.7754241867502432, "grad_norm": 0.4352587163448334, "learning_rate": 8e-05, "loss": 1.5789, "step": 7175 }, { "epoch": 0.77553225980763, "grad_norm": 0.3966296315193176, "learning_rate": 8e-05, "loss": 1.6816, "step": 7176 }, { "epoch": 0.7756403328650168, "grad_norm": 0.3940507769584656, "learning_rate": 8e-05, "loss": 1.5761, "step": 7177 }, { "epoch": 0.7757484059224036, "grad_norm": 0.47754427790641785, "learning_rate": 8e-05, "loss": 1.9542, "step": 7178 }, { "epoch": 0.7758564789797904, "grad_norm": 0.426866352558136, "learning_rate": 8e-05, "loss": 1.5223, "step": 7179 }, { "epoch": 0.7759645520371772, "grad_norm": 0.42209264636039734, "learning_rate": 8e-05, "loss": 1.8217, "step": 7180 }, { "epoch": 0.7760726250945639, "grad_norm": 0.41593310236930847, "learning_rate": 8e-05, "loss": 1.6922, "step": 7181 }, { "epoch": 0.7761806981519507, "grad_norm": 0.42591774463653564, "learning_rate": 8e-05, "loss": 1.6471, "step": 7182 }, { "epoch": 0.7762887712093375, "grad_norm": 0.4381764531135559, "learning_rate": 8e-05, "loss": 1.7448, "step": 7183 }, { "epoch": 0.7763968442667243, "grad_norm": 0.4404945969581604, "learning_rate": 8e-05, "loss": 1.5898, "step": 7184 }, { "epoch": 0.7765049173241111, "grad_norm": 0.40992873907089233, "learning_rate": 8e-05, "loss": 1.6641, "step": 7185 }, { "epoch": 0.7766129903814979, "grad_norm": 0.4530614912509918, "learning_rate": 8e-05, "loss": 1.7117, "step": 7186 }, { "epoch": 0.7767210634388847, "grad_norm": 0.43929627537727356, "learning_rate": 8e-05, "loss": 1.7017, "step": 7187 }, { "epoch": 0.7768291364962715, "grad_norm": 0.4431157112121582, "learning_rate": 8e-05, "loss": 1.8674, "step": 7188 }, { "epoch": 0.7769372095536583, "grad_norm": 0.4612402617931366, "learning_rate": 8e-05, "loss": 1.7345, "step": 7189 }, { "epoch": 0.7770452826110451, "grad_norm": 0.4507140517234802, "learning_rate": 8e-05, "loss": 1.4063, "step": 7190 }, { "epoch": 0.7771533556684319, "grad_norm": 0.42918336391448975, "learning_rate": 8e-05, "loss": 1.7936, "step": 7191 }, { "epoch": 0.7772614287258186, "grad_norm": 0.43238842487335205, "learning_rate": 8e-05, "loss": 1.5991, "step": 7192 }, { "epoch": 0.7773695017832054, "grad_norm": 0.5758771896362305, "learning_rate": 8e-05, "loss": 1.8455, "step": 7193 }, { "epoch": 0.7774775748405922, "grad_norm": 0.42827537655830383, "learning_rate": 8e-05, "loss": 1.4686, "step": 7194 }, { "epoch": 0.777585647897979, "grad_norm": 0.4057382047176361, "learning_rate": 8e-05, "loss": 1.5737, "step": 7195 }, { "epoch": 0.7776937209553658, "grad_norm": 0.47532251477241516, "learning_rate": 8e-05, "loss": 1.7449, "step": 7196 }, { "epoch": 0.7778017940127526, "grad_norm": 0.4262605905532837, "learning_rate": 8e-05, "loss": 1.7066, "step": 7197 }, { "epoch": 0.7779098670701394, "grad_norm": 0.4195728302001953, "learning_rate": 8e-05, "loss": 1.4374, "step": 7198 }, { "epoch": 0.7780179401275262, "grad_norm": 0.41859570145606995, "learning_rate": 8e-05, "loss": 1.6605, "step": 7199 }, { "epoch": 0.778126013184913, "grad_norm": 0.47693872451782227, "learning_rate": 8e-05, "loss": 1.727, "step": 7200 }, { "epoch": 0.7782340862422998, "grad_norm": 0.4388328492641449, "learning_rate": 8e-05, "loss": 1.5316, "step": 7201 }, { "epoch": 0.7783421592996866, "grad_norm": 0.39918479323387146, "learning_rate": 8e-05, "loss": 1.6551, "step": 7202 }, { "epoch": 0.7784502323570733, "grad_norm": 0.4378868043422699, "learning_rate": 8e-05, "loss": 1.6243, "step": 7203 }, { "epoch": 0.7785583054144601, "grad_norm": 0.4234863519668579, "learning_rate": 8e-05, "loss": 1.7047, "step": 7204 }, { "epoch": 0.7786663784718469, "grad_norm": 0.41522908210754395, "learning_rate": 8e-05, "loss": 1.6604, "step": 7205 }, { "epoch": 0.7787744515292337, "grad_norm": 0.4487966001033783, "learning_rate": 8e-05, "loss": 1.6782, "step": 7206 }, { "epoch": 0.7788825245866206, "grad_norm": 0.4448980987071991, "learning_rate": 8e-05, "loss": 1.7202, "step": 7207 }, { "epoch": 0.7789905976440074, "grad_norm": 0.40957313776016235, "learning_rate": 8e-05, "loss": 1.6381, "step": 7208 }, { "epoch": 0.7790986707013942, "grad_norm": 0.3983897864818573, "learning_rate": 8e-05, "loss": 1.6216, "step": 7209 }, { "epoch": 0.779206743758781, "grad_norm": 0.42147305607795715, "learning_rate": 8e-05, "loss": 1.6593, "step": 7210 }, { "epoch": 0.7793148168161678, "grad_norm": 0.48768776655197144, "learning_rate": 8e-05, "loss": 1.7807, "step": 7211 }, { "epoch": 0.7794228898735546, "grad_norm": 0.4381324350833893, "learning_rate": 8e-05, "loss": 1.6544, "step": 7212 }, { "epoch": 0.7795309629309413, "grad_norm": 0.42803293466567993, "learning_rate": 8e-05, "loss": 1.726, "step": 7213 }, { "epoch": 0.7796390359883281, "grad_norm": 0.441150963306427, "learning_rate": 8e-05, "loss": 1.4999, "step": 7214 }, { "epoch": 0.7797471090457149, "grad_norm": 0.4329712688922882, "learning_rate": 8e-05, "loss": 1.6594, "step": 7215 }, { "epoch": 0.7798551821031017, "grad_norm": 0.43992799520492554, "learning_rate": 8e-05, "loss": 1.7083, "step": 7216 }, { "epoch": 0.7799632551604885, "grad_norm": 0.4731760621070862, "learning_rate": 8e-05, "loss": 1.681, "step": 7217 }, { "epoch": 0.7800713282178753, "grad_norm": 0.44419172406196594, "learning_rate": 8e-05, "loss": 1.7308, "step": 7218 }, { "epoch": 0.7801794012752621, "grad_norm": 0.4423988461494446, "learning_rate": 8e-05, "loss": 1.6858, "step": 7219 }, { "epoch": 0.7802874743326489, "grad_norm": 0.41163700819015503, "learning_rate": 8e-05, "loss": 1.6022, "step": 7220 }, { "epoch": 0.7803955473900357, "grad_norm": 0.4506268799304962, "learning_rate": 8e-05, "loss": 1.641, "step": 7221 }, { "epoch": 0.7805036204474225, "grad_norm": 0.48903346061706543, "learning_rate": 8e-05, "loss": 1.7764, "step": 7222 }, { "epoch": 0.7806116935048093, "grad_norm": 0.4385514557361603, "learning_rate": 8e-05, "loss": 1.4645, "step": 7223 }, { "epoch": 0.780719766562196, "grad_norm": 0.5616792440414429, "learning_rate": 8e-05, "loss": 1.9961, "step": 7224 }, { "epoch": 0.7808278396195828, "grad_norm": 0.4503301978111267, "learning_rate": 8e-05, "loss": 1.7223, "step": 7225 }, { "epoch": 0.7809359126769696, "grad_norm": 0.46423399448394775, "learning_rate": 8e-05, "loss": 1.6099, "step": 7226 }, { "epoch": 0.7810439857343564, "grad_norm": 0.5086260437965393, "learning_rate": 8e-05, "loss": 1.6137, "step": 7227 }, { "epoch": 0.7811520587917432, "grad_norm": 0.5369820594787598, "learning_rate": 8e-05, "loss": 1.7828, "step": 7228 }, { "epoch": 0.78126013184913, "grad_norm": 0.44388526678085327, "learning_rate": 8e-05, "loss": 1.6985, "step": 7229 }, { "epoch": 0.7813682049065168, "grad_norm": 0.43774139881134033, "learning_rate": 8e-05, "loss": 1.6481, "step": 7230 }, { "epoch": 0.7814762779639036, "grad_norm": 0.44570156931877136, "learning_rate": 8e-05, "loss": 1.7066, "step": 7231 }, { "epoch": 0.7815843510212904, "grad_norm": 0.44153228402137756, "learning_rate": 8e-05, "loss": 1.7136, "step": 7232 }, { "epoch": 0.7816924240786772, "grad_norm": 0.43414992094039917, "learning_rate": 8e-05, "loss": 1.6863, "step": 7233 }, { "epoch": 0.781800497136064, "grad_norm": 0.4540584087371826, "learning_rate": 8e-05, "loss": 1.5915, "step": 7234 }, { "epoch": 0.7819085701934507, "grad_norm": 0.45434442162513733, "learning_rate": 8e-05, "loss": 1.6022, "step": 7235 }, { "epoch": 0.7820166432508375, "grad_norm": 0.40779924392700195, "learning_rate": 8e-05, "loss": 1.7104, "step": 7236 }, { "epoch": 0.7821247163082243, "grad_norm": 0.41777312755584717, "learning_rate": 8e-05, "loss": 1.5973, "step": 7237 }, { "epoch": 0.7822327893656111, "grad_norm": 0.4218105971813202, "learning_rate": 8e-05, "loss": 1.6544, "step": 7238 }, { "epoch": 0.7823408624229979, "grad_norm": 0.4527086913585663, "learning_rate": 8e-05, "loss": 1.6031, "step": 7239 }, { "epoch": 0.7824489354803847, "grad_norm": 0.4297373592853546, "learning_rate": 8e-05, "loss": 1.6849, "step": 7240 }, { "epoch": 0.7825570085377715, "grad_norm": 0.4250124394893646, "learning_rate": 8e-05, "loss": 1.6369, "step": 7241 }, { "epoch": 0.7826650815951584, "grad_norm": 0.4128621816635132, "learning_rate": 8e-05, "loss": 1.6269, "step": 7242 }, { "epoch": 0.7827731546525452, "grad_norm": 0.43640097975730896, "learning_rate": 8e-05, "loss": 1.6562, "step": 7243 }, { "epoch": 0.782881227709932, "grad_norm": 0.44332021474838257, "learning_rate": 8e-05, "loss": 1.7494, "step": 7244 }, { "epoch": 0.7829893007673187, "grad_norm": 0.42141029238700867, "learning_rate": 8e-05, "loss": 1.7869, "step": 7245 }, { "epoch": 0.7830973738247055, "grad_norm": 0.41320130228996277, "learning_rate": 8e-05, "loss": 1.6037, "step": 7246 }, { "epoch": 0.7832054468820923, "grad_norm": 0.4319283664226532, "learning_rate": 8e-05, "loss": 1.7105, "step": 7247 }, { "epoch": 0.7833135199394791, "grad_norm": 0.42235255241394043, "learning_rate": 8e-05, "loss": 1.6066, "step": 7248 }, { "epoch": 0.7834215929968659, "grad_norm": 0.4292883574962616, "learning_rate": 8e-05, "loss": 1.5335, "step": 7249 }, { "epoch": 0.7835296660542527, "grad_norm": 0.431995689868927, "learning_rate": 8e-05, "loss": 1.786, "step": 7250 }, { "epoch": 0.7836377391116395, "grad_norm": 0.42758721113204956, "learning_rate": 8e-05, "loss": 1.6835, "step": 7251 }, { "epoch": 0.7837458121690263, "grad_norm": 0.4359610676765442, "learning_rate": 8e-05, "loss": 1.6366, "step": 7252 }, { "epoch": 0.7838538852264131, "grad_norm": 0.4892270267009735, "learning_rate": 8e-05, "loss": 1.956, "step": 7253 }, { "epoch": 0.7839619582837999, "grad_norm": 0.45063236355781555, "learning_rate": 8e-05, "loss": 1.7113, "step": 7254 }, { "epoch": 0.7840700313411867, "grad_norm": 0.45067304372787476, "learning_rate": 8e-05, "loss": 1.5904, "step": 7255 }, { "epoch": 0.7841781043985734, "grad_norm": 0.41078853607177734, "learning_rate": 8e-05, "loss": 1.6638, "step": 7256 }, { "epoch": 0.7842861774559602, "grad_norm": 0.4184025526046753, "learning_rate": 8e-05, "loss": 1.535, "step": 7257 }, { "epoch": 0.784394250513347, "grad_norm": 0.435560941696167, "learning_rate": 8e-05, "loss": 1.7167, "step": 7258 }, { "epoch": 0.7845023235707338, "grad_norm": 0.4100242555141449, "learning_rate": 8e-05, "loss": 1.7314, "step": 7259 }, { "epoch": 0.7846103966281206, "grad_norm": 0.4492865204811096, "learning_rate": 8e-05, "loss": 1.7603, "step": 7260 }, { "epoch": 0.7847184696855074, "grad_norm": 0.4680573642253876, "learning_rate": 8e-05, "loss": 1.7932, "step": 7261 }, { "epoch": 0.7848265427428942, "grad_norm": 0.42632588744163513, "learning_rate": 8e-05, "loss": 1.7036, "step": 7262 }, { "epoch": 0.784934615800281, "grad_norm": 0.39122268557548523, "learning_rate": 8e-05, "loss": 1.5212, "step": 7263 }, { "epoch": 0.7850426888576678, "grad_norm": 0.4027564227581024, "learning_rate": 8e-05, "loss": 1.6987, "step": 7264 }, { "epoch": 0.7851507619150546, "grad_norm": 0.4085686504840851, "learning_rate": 8e-05, "loss": 1.7763, "step": 7265 }, { "epoch": 0.7852588349724414, "grad_norm": 0.43049725890159607, "learning_rate": 8e-05, "loss": 1.6131, "step": 7266 }, { "epoch": 0.7853669080298281, "grad_norm": 0.45872440934181213, "learning_rate": 8e-05, "loss": 1.7361, "step": 7267 }, { "epoch": 0.7854749810872149, "grad_norm": 0.4381159842014313, "learning_rate": 8e-05, "loss": 1.6611, "step": 7268 }, { "epoch": 0.7855830541446017, "grad_norm": 0.42267853021621704, "learning_rate": 8e-05, "loss": 1.691, "step": 7269 }, { "epoch": 0.7856911272019885, "grad_norm": 0.4073264002799988, "learning_rate": 8e-05, "loss": 1.6818, "step": 7270 }, { "epoch": 0.7857992002593753, "grad_norm": 0.4249771237373352, "learning_rate": 8e-05, "loss": 1.693, "step": 7271 }, { "epoch": 0.7859072733167621, "grad_norm": 0.44571003317832947, "learning_rate": 8e-05, "loss": 1.8258, "step": 7272 }, { "epoch": 0.7860153463741489, "grad_norm": 0.43842849135398865, "learning_rate": 8e-05, "loss": 1.7461, "step": 7273 }, { "epoch": 0.7861234194315357, "grad_norm": 0.4298919141292572, "learning_rate": 8e-05, "loss": 1.6432, "step": 7274 }, { "epoch": 0.7862314924889225, "grad_norm": 0.4257732629776001, "learning_rate": 8e-05, "loss": 1.8007, "step": 7275 }, { "epoch": 0.7863395655463093, "grad_norm": 0.49186763167381287, "learning_rate": 8e-05, "loss": 1.8269, "step": 7276 }, { "epoch": 0.7864476386036962, "grad_norm": 0.49494168162345886, "learning_rate": 8e-05, "loss": 1.9273, "step": 7277 }, { "epoch": 0.7865557116610828, "grad_norm": 0.44249334931373596, "learning_rate": 8e-05, "loss": 1.6754, "step": 7278 }, { "epoch": 0.7866637847184696, "grad_norm": 0.4260905683040619, "learning_rate": 8e-05, "loss": 1.6377, "step": 7279 }, { "epoch": 0.7867718577758565, "grad_norm": 0.4143972396850586, "learning_rate": 8e-05, "loss": 1.5601, "step": 7280 }, { "epoch": 0.7868799308332433, "grad_norm": 0.4299936294555664, "learning_rate": 8e-05, "loss": 1.8502, "step": 7281 }, { "epoch": 0.7869880038906301, "grad_norm": 0.4115182161331177, "learning_rate": 8e-05, "loss": 1.64, "step": 7282 }, { "epoch": 0.7870960769480169, "grad_norm": 0.40779030323028564, "learning_rate": 8e-05, "loss": 1.6437, "step": 7283 }, { "epoch": 0.7872041500054037, "grad_norm": 0.42192304134368896, "learning_rate": 8e-05, "loss": 1.4553, "step": 7284 }, { "epoch": 0.7873122230627905, "grad_norm": 0.49475955963134766, "learning_rate": 8e-05, "loss": 1.8452, "step": 7285 }, { "epoch": 0.7874202961201773, "grad_norm": 0.4084111750125885, "learning_rate": 8e-05, "loss": 1.4552, "step": 7286 }, { "epoch": 0.7875283691775641, "grad_norm": 0.4274614155292511, "learning_rate": 8e-05, "loss": 1.7411, "step": 7287 }, { "epoch": 0.7876364422349508, "grad_norm": 0.48056653141975403, "learning_rate": 8e-05, "loss": 1.6135, "step": 7288 }, { "epoch": 0.7877445152923376, "grad_norm": 0.44973495602607727, "learning_rate": 8e-05, "loss": 1.8427, "step": 7289 }, { "epoch": 0.7878525883497244, "grad_norm": 0.4326809346675873, "learning_rate": 8e-05, "loss": 1.666, "step": 7290 }, { "epoch": 0.7879606614071112, "grad_norm": 0.41282305121421814, "learning_rate": 8e-05, "loss": 1.5514, "step": 7291 }, { "epoch": 0.788068734464498, "grad_norm": 0.4271543025970459, "learning_rate": 8e-05, "loss": 1.7215, "step": 7292 }, { "epoch": 0.7881768075218848, "grad_norm": 0.4702816307544708, "learning_rate": 8e-05, "loss": 1.493, "step": 7293 }, { "epoch": 0.7882848805792716, "grad_norm": 0.411081463098526, "learning_rate": 8e-05, "loss": 1.4768, "step": 7294 }, { "epoch": 0.7883929536366584, "grad_norm": 0.47955650091171265, "learning_rate": 8e-05, "loss": 1.672, "step": 7295 }, { "epoch": 0.7885010266940452, "grad_norm": 0.49394720792770386, "learning_rate": 8e-05, "loss": 1.7568, "step": 7296 }, { "epoch": 0.788609099751432, "grad_norm": 0.4171009361743927, "learning_rate": 8e-05, "loss": 1.7441, "step": 7297 }, { "epoch": 0.7887171728088188, "grad_norm": 0.4318629205226898, "learning_rate": 8e-05, "loss": 1.6761, "step": 7298 }, { "epoch": 0.7888252458662055, "grad_norm": 0.4373484253883362, "learning_rate": 8e-05, "loss": 1.7264, "step": 7299 }, { "epoch": 0.7889333189235923, "grad_norm": 0.4445180296897888, "learning_rate": 8e-05, "loss": 1.7195, "step": 7300 }, { "epoch": 0.7890413919809791, "grad_norm": 0.42372819781303406, "learning_rate": 8e-05, "loss": 1.652, "step": 7301 }, { "epoch": 0.7891494650383659, "grad_norm": 0.4236440360546112, "learning_rate": 8e-05, "loss": 1.6453, "step": 7302 }, { "epoch": 0.7892575380957527, "grad_norm": 0.4614328145980835, "learning_rate": 8e-05, "loss": 1.7002, "step": 7303 }, { "epoch": 0.7893656111531395, "grad_norm": 0.4211183190345764, "learning_rate": 8e-05, "loss": 1.6986, "step": 7304 }, { "epoch": 0.7894736842105263, "grad_norm": 0.48019134998321533, "learning_rate": 8e-05, "loss": 1.7488, "step": 7305 }, { "epoch": 0.7895817572679131, "grad_norm": 0.3984740674495697, "learning_rate": 8e-05, "loss": 1.5815, "step": 7306 }, { "epoch": 0.7896898303252999, "grad_norm": 0.42727601528167725, "learning_rate": 8e-05, "loss": 1.6971, "step": 7307 }, { "epoch": 0.7897979033826867, "grad_norm": 0.42566925287246704, "learning_rate": 8e-05, "loss": 1.5731, "step": 7308 }, { "epoch": 0.7899059764400735, "grad_norm": 0.42615747451782227, "learning_rate": 8e-05, "loss": 1.7187, "step": 7309 }, { "epoch": 0.7900140494974602, "grad_norm": 0.429153174161911, "learning_rate": 8e-05, "loss": 1.6099, "step": 7310 }, { "epoch": 0.790122122554847, "grad_norm": 0.4287281632423401, "learning_rate": 8e-05, "loss": 1.5536, "step": 7311 }, { "epoch": 0.7902301956122338, "grad_norm": 0.4458554983139038, "learning_rate": 8e-05, "loss": 1.6626, "step": 7312 }, { "epoch": 0.7903382686696206, "grad_norm": 0.45236390829086304, "learning_rate": 8e-05, "loss": 1.7223, "step": 7313 }, { "epoch": 0.7904463417270075, "grad_norm": 0.4000852704048157, "learning_rate": 8e-05, "loss": 1.448, "step": 7314 }, { "epoch": 0.7905544147843943, "grad_norm": 0.44914889335632324, "learning_rate": 8e-05, "loss": 1.647, "step": 7315 }, { "epoch": 0.7906624878417811, "grad_norm": 0.4580637812614441, "learning_rate": 8e-05, "loss": 1.7155, "step": 7316 }, { "epoch": 0.7907705608991679, "grad_norm": 0.5046891570091248, "learning_rate": 8e-05, "loss": 1.7536, "step": 7317 }, { "epoch": 0.7908786339565547, "grad_norm": 0.3993486166000366, "learning_rate": 8e-05, "loss": 1.5751, "step": 7318 }, { "epoch": 0.7909867070139415, "grad_norm": 0.44040340185165405, "learning_rate": 8e-05, "loss": 1.6057, "step": 7319 }, { "epoch": 0.7910947800713282, "grad_norm": 0.4745579659938812, "learning_rate": 8e-05, "loss": 1.8416, "step": 7320 }, { "epoch": 0.791202853128715, "grad_norm": 0.400669664144516, "learning_rate": 8e-05, "loss": 1.595, "step": 7321 }, { "epoch": 0.7913109261861018, "grad_norm": 0.4103098213672638, "learning_rate": 8e-05, "loss": 1.5181, "step": 7322 }, { "epoch": 0.7914189992434886, "grad_norm": 0.40182948112487793, "learning_rate": 8e-05, "loss": 1.5808, "step": 7323 }, { "epoch": 0.7915270723008754, "grad_norm": 0.44161295890808105, "learning_rate": 8e-05, "loss": 1.953, "step": 7324 }, { "epoch": 0.7916351453582622, "grad_norm": 0.43933144211769104, "learning_rate": 8e-05, "loss": 1.7329, "step": 7325 }, { "epoch": 0.791743218415649, "grad_norm": 0.45062634348869324, "learning_rate": 8e-05, "loss": 1.6186, "step": 7326 }, { "epoch": 0.7918512914730358, "grad_norm": 0.44705337285995483, "learning_rate": 8e-05, "loss": 1.6991, "step": 7327 }, { "epoch": 0.7919593645304226, "grad_norm": 0.44825002551078796, "learning_rate": 8e-05, "loss": 1.5222, "step": 7328 }, { "epoch": 0.7920674375878094, "grad_norm": 0.4779050946235657, "learning_rate": 8e-05, "loss": 1.7305, "step": 7329 }, { "epoch": 0.7921755106451962, "grad_norm": 0.4632016718387604, "learning_rate": 8e-05, "loss": 1.9141, "step": 7330 }, { "epoch": 0.7922835837025829, "grad_norm": 0.43903419375419617, "learning_rate": 8e-05, "loss": 1.6867, "step": 7331 }, { "epoch": 0.7923916567599697, "grad_norm": 0.39453765749931335, "learning_rate": 8e-05, "loss": 1.3289, "step": 7332 }, { "epoch": 0.7924997298173565, "grad_norm": 0.5118459463119507, "learning_rate": 8e-05, "loss": 1.7266, "step": 7333 }, { "epoch": 0.7926078028747433, "grad_norm": 0.43074822425842285, "learning_rate": 8e-05, "loss": 1.6926, "step": 7334 }, { "epoch": 0.7927158759321301, "grad_norm": 0.472474068403244, "learning_rate": 8e-05, "loss": 1.7455, "step": 7335 }, { "epoch": 0.7928239489895169, "grad_norm": 0.46575117111206055, "learning_rate": 8e-05, "loss": 1.7009, "step": 7336 }, { "epoch": 0.7929320220469037, "grad_norm": 0.45209363102912903, "learning_rate": 8e-05, "loss": 1.683, "step": 7337 }, { "epoch": 0.7930400951042905, "grad_norm": 0.4139028489589691, "learning_rate": 8e-05, "loss": 1.6358, "step": 7338 }, { "epoch": 0.7931481681616773, "grad_norm": 0.4461669623851776, "learning_rate": 8e-05, "loss": 1.6471, "step": 7339 }, { "epoch": 0.7932562412190641, "grad_norm": 0.44337955117225647, "learning_rate": 8e-05, "loss": 1.7385, "step": 7340 }, { "epoch": 0.7933643142764509, "grad_norm": 0.4264039993286133, "learning_rate": 8e-05, "loss": 1.6021, "step": 7341 }, { "epoch": 0.7934723873338376, "grad_norm": 0.43781912326812744, "learning_rate": 8e-05, "loss": 1.6886, "step": 7342 }, { "epoch": 0.7935804603912244, "grad_norm": 0.4672248661518097, "learning_rate": 8e-05, "loss": 1.7781, "step": 7343 }, { "epoch": 0.7936885334486112, "grad_norm": 0.460803359746933, "learning_rate": 8e-05, "loss": 1.7335, "step": 7344 }, { "epoch": 0.793796606505998, "grad_norm": 0.43807336688041687, "learning_rate": 8e-05, "loss": 1.6989, "step": 7345 }, { "epoch": 0.7939046795633848, "grad_norm": 0.5371410846710205, "learning_rate": 8e-05, "loss": 1.8704, "step": 7346 }, { "epoch": 0.7940127526207716, "grad_norm": 0.43442049622535706, "learning_rate": 8e-05, "loss": 1.7543, "step": 7347 }, { "epoch": 0.7941208256781584, "grad_norm": 0.4258841872215271, "learning_rate": 8e-05, "loss": 1.543, "step": 7348 }, { "epoch": 0.7942288987355453, "grad_norm": 0.4319019913673401, "learning_rate": 8e-05, "loss": 1.843, "step": 7349 }, { "epoch": 0.7943369717929321, "grad_norm": 0.4540032744407654, "learning_rate": 8e-05, "loss": 1.7847, "step": 7350 }, { "epoch": 0.7944450448503189, "grad_norm": 0.43536654114723206, "learning_rate": 8e-05, "loss": 1.7241, "step": 7351 }, { "epoch": 0.7945531179077057, "grad_norm": 0.44670623540878296, "learning_rate": 8e-05, "loss": 1.5429, "step": 7352 }, { "epoch": 0.7946611909650924, "grad_norm": 0.42692476511001587, "learning_rate": 8e-05, "loss": 1.653, "step": 7353 }, { "epoch": 0.7947692640224792, "grad_norm": 0.4593164622783661, "learning_rate": 8e-05, "loss": 1.6794, "step": 7354 }, { "epoch": 0.794877337079866, "grad_norm": 0.4186803102493286, "learning_rate": 8e-05, "loss": 1.7864, "step": 7355 }, { "epoch": 0.7949854101372528, "grad_norm": 0.42590057849884033, "learning_rate": 8e-05, "loss": 1.6899, "step": 7356 }, { "epoch": 0.7950934831946396, "grad_norm": 0.4129270017147064, "learning_rate": 8e-05, "loss": 1.6682, "step": 7357 }, { "epoch": 0.7952015562520264, "grad_norm": 0.4284450113773346, "learning_rate": 8e-05, "loss": 1.5976, "step": 7358 }, { "epoch": 0.7953096293094132, "grad_norm": 0.4026896357536316, "learning_rate": 8e-05, "loss": 1.6413, "step": 7359 }, { "epoch": 0.7954177023668, "grad_norm": 0.4129711985588074, "learning_rate": 8e-05, "loss": 1.6061, "step": 7360 }, { "epoch": 0.7955257754241868, "grad_norm": 0.4233400225639343, "learning_rate": 8e-05, "loss": 1.5908, "step": 7361 }, { "epoch": 0.7956338484815736, "grad_norm": 0.4437609016895294, "learning_rate": 8e-05, "loss": 1.6628, "step": 7362 }, { "epoch": 0.7957419215389603, "grad_norm": 0.421798974275589, "learning_rate": 8e-05, "loss": 1.7101, "step": 7363 }, { "epoch": 0.7958499945963471, "grad_norm": 0.41736412048339844, "learning_rate": 8e-05, "loss": 1.6526, "step": 7364 }, { "epoch": 0.7959580676537339, "grad_norm": 0.41071435809135437, "learning_rate": 8e-05, "loss": 1.5138, "step": 7365 }, { "epoch": 0.7960661407111207, "grad_norm": 0.4442670941352844, "learning_rate": 8e-05, "loss": 1.7246, "step": 7366 }, { "epoch": 0.7961742137685075, "grad_norm": 0.45115745067596436, "learning_rate": 8e-05, "loss": 1.7093, "step": 7367 }, { "epoch": 0.7962822868258943, "grad_norm": 0.43646007776260376, "learning_rate": 8e-05, "loss": 1.6045, "step": 7368 }, { "epoch": 0.7963903598832811, "grad_norm": 0.4515870213508606, "learning_rate": 8e-05, "loss": 1.7042, "step": 7369 }, { "epoch": 0.7964984329406679, "grad_norm": 0.5268057584762573, "learning_rate": 8e-05, "loss": 1.7648, "step": 7370 }, { "epoch": 0.7966065059980547, "grad_norm": 0.4220568537712097, "learning_rate": 8e-05, "loss": 1.6073, "step": 7371 }, { "epoch": 0.7967145790554415, "grad_norm": 0.48166730999946594, "learning_rate": 8e-05, "loss": 1.828, "step": 7372 }, { "epoch": 0.7968226521128283, "grad_norm": 0.415290892124176, "learning_rate": 8e-05, "loss": 1.6265, "step": 7373 }, { "epoch": 0.796930725170215, "grad_norm": 0.4048038721084595, "learning_rate": 8e-05, "loss": 1.5307, "step": 7374 }, { "epoch": 0.7970387982276018, "grad_norm": 0.40419334173202515, "learning_rate": 8e-05, "loss": 1.6331, "step": 7375 }, { "epoch": 0.7971468712849886, "grad_norm": 0.4365962743759155, "learning_rate": 8e-05, "loss": 1.7217, "step": 7376 }, { "epoch": 0.7972549443423754, "grad_norm": 0.43775978684425354, "learning_rate": 8e-05, "loss": 1.7619, "step": 7377 }, { "epoch": 0.7973630173997622, "grad_norm": 0.4677794575691223, "learning_rate": 8e-05, "loss": 1.7429, "step": 7378 }, { "epoch": 0.797471090457149, "grad_norm": 0.4253327548503876, "learning_rate": 8e-05, "loss": 1.6419, "step": 7379 }, { "epoch": 0.7975791635145358, "grad_norm": 0.41763684153556824, "learning_rate": 8e-05, "loss": 1.5659, "step": 7380 }, { "epoch": 0.7976872365719226, "grad_norm": 0.38707825541496277, "learning_rate": 8e-05, "loss": 1.5311, "step": 7381 }, { "epoch": 0.7977953096293094, "grad_norm": 0.3834069073200226, "learning_rate": 8e-05, "loss": 1.4477, "step": 7382 }, { "epoch": 0.7979033826866962, "grad_norm": 0.4635106027126312, "learning_rate": 8e-05, "loss": 1.6706, "step": 7383 }, { "epoch": 0.798011455744083, "grad_norm": 0.47511518001556396, "learning_rate": 8e-05, "loss": 1.7122, "step": 7384 }, { "epoch": 0.7981195288014697, "grad_norm": 0.5376856327056885, "learning_rate": 8e-05, "loss": 2.1139, "step": 7385 }, { "epoch": 0.7982276018588566, "grad_norm": 0.47975051403045654, "learning_rate": 8e-05, "loss": 1.7137, "step": 7386 }, { "epoch": 0.7983356749162434, "grad_norm": 0.3920920193195343, "learning_rate": 8e-05, "loss": 1.4125, "step": 7387 }, { "epoch": 0.7984437479736302, "grad_norm": 0.41979706287384033, "learning_rate": 8e-05, "loss": 1.7063, "step": 7388 }, { "epoch": 0.798551821031017, "grad_norm": 0.5022977590560913, "learning_rate": 8e-05, "loss": 1.9152, "step": 7389 }, { "epoch": 0.7986598940884038, "grad_norm": 0.42217475175857544, "learning_rate": 8e-05, "loss": 1.5725, "step": 7390 }, { "epoch": 0.7987679671457906, "grad_norm": 0.43227124214172363, "learning_rate": 8e-05, "loss": 1.681, "step": 7391 }, { "epoch": 0.7988760402031774, "grad_norm": 0.4352397322654724, "learning_rate": 8e-05, "loss": 1.7454, "step": 7392 }, { "epoch": 0.7989841132605642, "grad_norm": 0.4385949373245239, "learning_rate": 8e-05, "loss": 1.628, "step": 7393 }, { "epoch": 0.799092186317951, "grad_norm": 0.45685943961143494, "learning_rate": 8e-05, "loss": 1.6815, "step": 7394 }, { "epoch": 0.7992002593753377, "grad_norm": 0.4925706088542938, "learning_rate": 8e-05, "loss": 1.7275, "step": 7395 }, { "epoch": 0.7993083324327245, "grad_norm": 0.45838430523872375, "learning_rate": 8e-05, "loss": 1.739, "step": 7396 }, { "epoch": 0.7994164054901113, "grad_norm": 0.43183228373527527, "learning_rate": 8e-05, "loss": 1.57, "step": 7397 }, { "epoch": 0.7995244785474981, "grad_norm": 0.4356153905391693, "learning_rate": 8e-05, "loss": 1.6188, "step": 7398 }, { "epoch": 0.7996325516048849, "grad_norm": 0.42947453260421753, "learning_rate": 8e-05, "loss": 1.6589, "step": 7399 }, { "epoch": 0.7997406246622717, "grad_norm": 0.4540838301181793, "learning_rate": 8e-05, "loss": 1.6517, "step": 7400 }, { "epoch": 0.7998486977196585, "grad_norm": 0.42616090178489685, "learning_rate": 8e-05, "loss": 1.6396, "step": 7401 }, { "epoch": 0.7999567707770453, "grad_norm": 0.44712892174720764, "learning_rate": 8e-05, "loss": 1.8332, "step": 7402 }, { "epoch": 0.8000648438344321, "grad_norm": 0.48440051078796387, "learning_rate": 8e-05, "loss": 1.6484, "step": 7403 }, { "epoch": 0.8001729168918189, "grad_norm": 0.43263953924179077, "learning_rate": 8e-05, "loss": 1.6847, "step": 7404 }, { "epoch": 0.8002809899492057, "grad_norm": 0.40127867460250854, "learning_rate": 8e-05, "loss": 1.648, "step": 7405 }, { "epoch": 0.8003890630065924, "grad_norm": 0.41783565282821655, "learning_rate": 8e-05, "loss": 1.7268, "step": 7406 }, { "epoch": 0.8004971360639792, "grad_norm": 0.4405827522277832, "learning_rate": 8e-05, "loss": 1.658, "step": 7407 }, { "epoch": 0.800605209121366, "grad_norm": 0.4956062138080597, "learning_rate": 8e-05, "loss": 1.6093, "step": 7408 }, { "epoch": 0.8007132821787528, "grad_norm": 0.437845915555954, "learning_rate": 8e-05, "loss": 1.5189, "step": 7409 }, { "epoch": 0.8008213552361396, "grad_norm": 0.4335143268108368, "learning_rate": 8e-05, "loss": 1.604, "step": 7410 }, { "epoch": 0.8009294282935264, "grad_norm": 0.44935545325279236, "learning_rate": 8e-05, "loss": 1.8389, "step": 7411 }, { "epoch": 0.8010375013509132, "grad_norm": 0.4280940890312195, "learning_rate": 8e-05, "loss": 1.6409, "step": 7412 }, { "epoch": 0.8011455744083, "grad_norm": 0.40842336416244507, "learning_rate": 8e-05, "loss": 1.6269, "step": 7413 }, { "epoch": 0.8012536474656868, "grad_norm": 0.4249597191810608, "learning_rate": 8e-05, "loss": 1.6336, "step": 7414 }, { "epoch": 0.8013617205230736, "grad_norm": 0.41606390476226807, "learning_rate": 8e-05, "loss": 1.6626, "step": 7415 }, { "epoch": 0.8014697935804604, "grad_norm": 0.4775324761867523, "learning_rate": 8e-05, "loss": 1.797, "step": 7416 }, { "epoch": 0.8015778666378471, "grad_norm": 0.40998759865760803, "learning_rate": 8e-05, "loss": 1.4899, "step": 7417 }, { "epoch": 0.8016859396952339, "grad_norm": 0.46148842573165894, "learning_rate": 8e-05, "loss": 1.7476, "step": 7418 }, { "epoch": 0.8017940127526207, "grad_norm": 0.4382099509239197, "learning_rate": 8e-05, "loss": 1.6651, "step": 7419 }, { "epoch": 0.8019020858100075, "grad_norm": 0.47996577620506287, "learning_rate": 8e-05, "loss": 1.4107, "step": 7420 }, { "epoch": 0.8020101588673944, "grad_norm": 0.4088434875011444, "learning_rate": 8e-05, "loss": 1.4061, "step": 7421 }, { "epoch": 0.8021182319247812, "grad_norm": 0.4736614525318146, "learning_rate": 8e-05, "loss": 1.7031, "step": 7422 }, { "epoch": 0.802226304982168, "grad_norm": 0.4343092143535614, "learning_rate": 8e-05, "loss": 1.7842, "step": 7423 }, { "epoch": 0.8023343780395548, "grad_norm": 0.43101316690444946, "learning_rate": 8e-05, "loss": 1.6901, "step": 7424 }, { "epoch": 0.8024424510969416, "grad_norm": 0.507854163646698, "learning_rate": 8e-05, "loss": 1.9718, "step": 7425 }, { "epoch": 0.8025505241543284, "grad_norm": 0.4384128451347351, "learning_rate": 8e-05, "loss": 1.5699, "step": 7426 }, { "epoch": 0.8026585972117152, "grad_norm": 0.41657692193984985, "learning_rate": 8e-05, "loss": 1.5461, "step": 7427 }, { "epoch": 0.8027666702691019, "grad_norm": 0.4214479923248291, "learning_rate": 8e-05, "loss": 1.6848, "step": 7428 }, { "epoch": 0.8028747433264887, "grad_norm": 0.44118770956993103, "learning_rate": 8e-05, "loss": 1.6546, "step": 7429 }, { "epoch": 0.8029828163838755, "grad_norm": 0.4571443200111389, "learning_rate": 8e-05, "loss": 1.6507, "step": 7430 }, { "epoch": 0.8030908894412623, "grad_norm": 0.465093731880188, "learning_rate": 8e-05, "loss": 1.8575, "step": 7431 }, { "epoch": 0.8031989624986491, "grad_norm": 0.46236562728881836, "learning_rate": 8e-05, "loss": 1.7445, "step": 7432 }, { "epoch": 0.8033070355560359, "grad_norm": 0.4298613965511322, "learning_rate": 8e-05, "loss": 1.657, "step": 7433 }, { "epoch": 0.8034151086134227, "grad_norm": 0.4285863935947418, "learning_rate": 8e-05, "loss": 1.3856, "step": 7434 }, { "epoch": 0.8035231816708095, "grad_norm": 0.45419952273368835, "learning_rate": 8e-05, "loss": 1.5867, "step": 7435 }, { "epoch": 0.8036312547281963, "grad_norm": 0.4437437355518341, "learning_rate": 8e-05, "loss": 1.8208, "step": 7436 }, { "epoch": 0.8037393277855831, "grad_norm": 0.45749640464782715, "learning_rate": 8e-05, "loss": 1.7741, "step": 7437 }, { "epoch": 0.8038474008429698, "grad_norm": 0.4751962125301361, "learning_rate": 8e-05, "loss": 1.9305, "step": 7438 }, { "epoch": 0.8039554739003566, "grad_norm": 0.44853436946868896, "learning_rate": 8e-05, "loss": 1.7834, "step": 7439 }, { "epoch": 0.8040635469577434, "grad_norm": 0.4352447986602783, "learning_rate": 8e-05, "loss": 1.5782, "step": 7440 }, { "epoch": 0.8041716200151302, "grad_norm": 0.4318731725215912, "learning_rate": 8e-05, "loss": 1.748, "step": 7441 }, { "epoch": 0.804279693072517, "grad_norm": 0.4502854645252228, "learning_rate": 8e-05, "loss": 1.6742, "step": 7442 }, { "epoch": 0.8043877661299038, "grad_norm": 0.5288558006286621, "learning_rate": 8e-05, "loss": 1.728, "step": 7443 }, { "epoch": 0.8044958391872906, "grad_norm": 0.4394971430301666, "learning_rate": 8e-05, "loss": 1.6564, "step": 7444 }, { "epoch": 0.8046039122446774, "grad_norm": 0.4773208200931549, "learning_rate": 8e-05, "loss": 1.7447, "step": 7445 }, { "epoch": 0.8047119853020642, "grad_norm": 0.4158342480659485, "learning_rate": 8e-05, "loss": 1.7139, "step": 7446 }, { "epoch": 0.804820058359451, "grad_norm": 0.43002575635910034, "learning_rate": 8e-05, "loss": 1.7011, "step": 7447 }, { "epoch": 0.8049281314168378, "grad_norm": 0.42534738779067993, "learning_rate": 8e-05, "loss": 1.7655, "step": 7448 }, { "epoch": 0.8050362044742245, "grad_norm": 0.4661007523536682, "learning_rate": 8e-05, "loss": 1.7576, "step": 7449 }, { "epoch": 0.8051442775316113, "grad_norm": 0.4557751715183258, "learning_rate": 8e-05, "loss": 1.7435, "step": 7450 }, { "epoch": 0.8052523505889981, "grad_norm": 0.46113595366477966, "learning_rate": 8e-05, "loss": 1.8264, "step": 7451 }, { "epoch": 0.8053604236463849, "grad_norm": 0.4181073307991028, "learning_rate": 8e-05, "loss": 1.5966, "step": 7452 }, { "epoch": 0.8054684967037717, "grad_norm": 0.4319812059402466, "learning_rate": 8e-05, "loss": 1.5873, "step": 7453 }, { "epoch": 0.8055765697611585, "grad_norm": 0.45194804668426514, "learning_rate": 8e-05, "loss": 1.7788, "step": 7454 }, { "epoch": 0.8056846428185453, "grad_norm": 0.4407847225666046, "learning_rate": 8e-05, "loss": 1.8215, "step": 7455 }, { "epoch": 0.8057927158759322, "grad_norm": 0.4888278543949127, "learning_rate": 8e-05, "loss": 1.8193, "step": 7456 }, { "epoch": 0.805900788933319, "grad_norm": 0.43444138765335083, "learning_rate": 8e-05, "loss": 1.7184, "step": 7457 }, { "epoch": 0.8060088619907058, "grad_norm": 0.4897560775279999, "learning_rate": 8e-05, "loss": 1.7543, "step": 7458 }, { "epoch": 0.8061169350480926, "grad_norm": 0.46517348289489746, "learning_rate": 8e-05, "loss": 1.8861, "step": 7459 }, { "epoch": 0.8062250081054793, "grad_norm": 0.4678691327571869, "learning_rate": 8e-05, "loss": 1.7819, "step": 7460 }, { "epoch": 0.8063330811628661, "grad_norm": 0.417611300945282, "learning_rate": 8e-05, "loss": 1.6817, "step": 7461 }, { "epoch": 0.8064411542202529, "grad_norm": 0.4068208336830139, "learning_rate": 8e-05, "loss": 1.5777, "step": 7462 }, { "epoch": 0.8065492272776397, "grad_norm": 0.471172034740448, "learning_rate": 8e-05, "loss": 1.7084, "step": 7463 }, { "epoch": 0.8066573003350265, "grad_norm": 0.4585971236228943, "learning_rate": 8e-05, "loss": 1.6761, "step": 7464 }, { "epoch": 0.8067653733924133, "grad_norm": 0.4551122188568115, "learning_rate": 8e-05, "loss": 1.8962, "step": 7465 }, { "epoch": 0.8068734464498001, "grad_norm": 0.4044235944747925, "learning_rate": 8e-05, "loss": 1.4993, "step": 7466 }, { "epoch": 0.8069815195071869, "grad_norm": 0.4167936146259308, "learning_rate": 8e-05, "loss": 1.7243, "step": 7467 }, { "epoch": 0.8070895925645737, "grad_norm": 0.42447394132614136, "learning_rate": 8e-05, "loss": 1.6239, "step": 7468 }, { "epoch": 0.8071976656219605, "grad_norm": 0.46412742137908936, "learning_rate": 8e-05, "loss": 1.7796, "step": 7469 }, { "epoch": 0.8073057386793472, "grad_norm": 0.5221142768859863, "learning_rate": 8e-05, "loss": 1.8337, "step": 7470 }, { "epoch": 0.807413811736734, "grad_norm": 0.4611998200416565, "learning_rate": 8e-05, "loss": 1.7308, "step": 7471 }, { "epoch": 0.8075218847941208, "grad_norm": 0.448544442653656, "learning_rate": 8e-05, "loss": 1.8104, "step": 7472 }, { "epoch": 0.8076299578515076, "grad_norm": 0.4568692743778229, "learning_rate": 8e-05, "loss": 1.714, "step": 7473 }, { "epoch": 0.8077380309088944, "grad_norm": 0.4173417389392853, "learning_rate": 8e-05, "loss": 1.7391, "step": 7474 }, { "epoch": 0.8078461039662812, "grad_norm": 0.4618779718875885, "learning_rate": 8e-05, "loss": 1.7972, "step": 7475 }, { "epoch": 0.807954177023668, "grad_norm": 0.41760188341140747, "learning_rate": 8e-05, "loss": 1.696, "step": 7476 }, { "epoch": 0.8080622500810548, "grad_norm": 0.4396481215953827, "learning_rate": 8e-05, "loss": 1.7436, "step": 7477 }, { "epoch": 0.8081703231384416, "grad_norm": 0.4402555525302887, "learning_rate": 8e-05, "loss": 1.7707, "step": 7478 }, { "epoch": 0.8082783961958284, "grad_norm": 0.4265132248401642, "learning_rate": 8e-05, "loss": 1.5679, "step": 7479 }, { "epoch": 0.8083864692532152, "grad_norm": 0.4388149082660675, "learning_rate": 8e-05, "loss": 1.8208, "step": 7480 }, { "epoch": 0.8084945423106019, "grad_norm": 0.4575382471084595, "learning_rate": 8e-05, "loss": 1.7625, "step": 7481 }, { "epoch": 0.8086026153679887, "grad_norm": 0.4178367853164673, "learning_rate": 8e-05, "loss": 1.562, "step": 7482 }, { "epoch": 0.8087106884253755, "grad_norm": 0.5554490089416504, "learning_rate": 8e-05, "loss": 1.7634, "step": 7483 }, { "epoch": 0.8088187614827623, "grad_norm": 0.44815313816070557, "learning_rate": 8e-05, "loss": 1.713, "step": 7484 }, { "epoch": 0.8089268345401491, "grad_norm": 0.5138227939605713, "learning_rate": 8e-05, "loss": 1.4113, "step": 7485 }, { "epoch": 0.8090349075975359, "grad_norm": 0.4367561638355255, "learning_rate": 8e-05, "loss": 1.7257, "step": 7486 }, { "epoch": 0.8091429806549227, "grad_norm": 0.4244377315044403, "learning_rate": 8e-05, "loss": 1.6909, "step": 7487 }, { "epoch": 0.8092510537123095, "grad_norm": 0.41467106342315674, "learning_rate": 8e-05, "loss": 1.7069, "step": 7488 }, { "epoch": 0.8093591267696963, "grad_norm": 0.48617467284202576, "learning_rate": 8e-05, "loss": 1.8346, "step": 7489 }, { "epoch": 0.8094671998270832, "grad_norm": 0.45712095499038696, "learning_rate": 8e-05, "loss": 1.5375, "step": 7490 }, { "epoch": 0.80957527288447, "grad_norm": 0.5039635896682739, "learning_rate": 8e-05, "loss": 1.8245, "step": 7491 }, { "epoch": 0.8096833459418566, "grad_norm": 0.42388734221458435, "learning_rate": 8e-05, "loss": 1.5649, "step": 7492 }, { "epoch": 0.8097914189992435, "grad_norm": 0.4325251877307892, "learning_rate": 8e-05, "loss": 1.5652, "step": 7493 }, { "epoch": 0.8098994920566303, "grad_norm": 0.4374862611293793, "learning_rate": 8e-05, "loss": 1.7439, "step": 7494 }, { "epoch": 0.8100075651140171, "grad_norm": 0.4471244513988495, "learning_rate": 8e-05, "loss": 1.7515, "step": 7495 }, { "epoch": 0.8101156381714039, "grad_norm": 0.463533878326416, "learning_rate": 8e-05, "loss": 1.6113, "step": 7496 }, { "epoch": 0.8102237112287907, "grad_norm": 0.4889056980609894, "learning_rate": 8e-05, "loss": 1.8742, "step": 7497 }, { "epoch": 0.8103317842861775, "grad_norm": 0.49760186672210693, "learning_rate": 8e-05, "loss": 1.6732, "step": 7498 }, { "epoch": 0.8104398573435643, "grad_norm": 0.47731876373291016, "learning_rate": 8e-05, "loss": 1.8357, "step": 7499 }, { "epoch": 0.8105479304009511, "grad_norm": 0.4264635443687439, "learning_rate": 8e-05, "loss": 1.6394, "step": 7500 }, { "epoch": 0.8106560034583379, "grad_norm": 0.447874516248703, "learning_rate": 8e-05, "loss": 1.7327, "step": 7501 }, { "epoch": 0.8107640765157246, "grad_norm": 0.460968554019928, "learning_rate": 8e-05, "loss": 1.7644, "step": 7502 }, { "epoch": 0.8108721495731114, "grad_norm": 0.423146516084671, "learning_rate": 8e-05, "loss": 1.6296, "step": 7503 }, { "epoch": 0.8109802226304982, "grad_norm": 0.4018899202346802, "learning_rate": 8e-05, "loss": 1.4135, "step": 7504 }, { "epoch": 0.811088295687885, "grad_norm": 0.4340698719024658, "learning_rate": 8e-05, "loss": 1.6397, "step": 7505 }, { "epoch": 0.8111963687452718, "grad_norm": 0.4268712103366852, "learning_rate": 8e-05, "loss": 1.6272, "step": 7506 }, { "epoch": 0.8113044418026586, "grad_norm": 0.44992977380752563, "learning_rate": 8e-05, "loss": 1.7023, "step": 7507 }, { "epoch": 0.8114125148600454, "grad_norm": 0.4349914789199829, "learning_rate": 8e-05, "loss": 1.6712, "step": 7508 }, { "epoch": 0.8115205879174322, "grad_norm": 0.4133839011192322, "learning_rate": 8e-05, "loss": 1.5972, "step": 7509 }, { "epoch": 0.811628660974819, "grad_norm": 0.41420769691467285, "learning_rate": 8e-05, "loss": 1.7688, "step": 7510 }, { "epoch": 0.8117367340322058, "grad_norm": 0.43037259578704834, "learning_rate": 8e-05, "loss": 1.621, "step": 7511 }, { "epoch": 0.8118448070895926, "grad_norm": 0.435275137424469, "learning_rate": 8e-05, "loss": 1.7305, "step": 7512 }, { "epoch": 0.8119528801469793, "grad_norm": 0.47098442912101746, "learning_rate": 8e-05, "loss": 1.5022, "step": 7513 }, { "epoch": 0.8120609532043661, "grad_norm": 0.4217282235622406, "learning_rate": 8e-05, "loss": 1.5125, "step": 7514 }, { "epoch": 0.8121690262617529, "grad_norm": 0.4228089451789856, "learning_rate": 8e-05, "loss": 1.6484, "step": 7515 }, { "epoch": 0.8122770993191397, "grad_norm": 0.45302578806877136, "learning_rate": 8e-05, "loss": 1.8631, "step": 7516 }, { "epoch": 0.8123851723765265, "grad_norm": 0.42124029994010925, "learning_rate": 8e-05, "loss": 1.7302, "step": 7517 }, { "epoch": 0.8124932454339133, "grad_norm": 0.4276401400566101, "learning_rate": 8e-05, "loss": 1.5925, "step": 7518 }, { "epoch": 0.8126013184913001, "grad_norm": 0.4372789263725281, "learning_rate": 8e-05, "loss": 1.5508, "step": 7519 }, { "epoch": 0.8127093915486869, "grad_norm": 0.5060647130012512, "learning_rate": 8e-05, "loss": 1.6306, "step": 7520 }, { "epoch": 0.8128174646060737, "grad_norm": 0.43591371178627014, "learning_rate": 8e-05, "loss": 1.7092, "step": 7521 }, { "epoch": 0.8129255376634605, "grad_norm": 0.45248350501060486, "learning_rate": 8e-05, "loss": 1.6709, "step": 7522 }, { "epoch": 0.8130336107208473, "grad_norm": 0.43095681071281433, "learning_rate": 8e-05, "loss": 1.6691, "step": 7523 }, { "epoch": 0.813141683778234, "grad_norm": 0.47643744945526123, "learning_rate": 8e-05, "loss": 1.7687, "step": 7524 }, { "epoch": 0.8132497568356208, "grad_norm": 0.41327545046806335, "learning_rate": 8e-05, "loss": 1.6047, "step": 7525 }, { "epoch": 0.8133578298930076, "grad_norm": 0.5220661759376526, "learning_rate": 8e-05, "loss": 1.8386, "step": 7526 }, { "epoch": 0.8134659029503944, "grad_norm": 0.4977021813392639, "learning_rate": 8e-05, "loss": 1.9418, "step": 7527 }, { "epoch": 0.8135739760077813, "grad_norm": 0.46191877126693726, "learning_rate": 8e-05, "loss": 1.7054, "step": 7528 }, { "epoch": 0.8136820490651681, "grad_norm": 0.4919244945049286, "learning_rate": 8e-05, "loss": 1.8003, "step": 7529 }, { "epoch": 0.8137901221225549, "grad_norm": 0.4270123839378357, "learning_rate": 8e-05, "loss": 1.7063, "step": 7530 }, { "epoch": 0.8138981951799417, "grad_norm": 0.46366336941719055, "learning_rate": 8e-05, "loss": 1.5868, "step": 7531 }, { "epoch": 0.8140062682373285, "grad_norm": 0.46643903851509094, "learning_rate": 8e-05, "loss": 1.7798, "step": 7532 }, { "epoch": 0.8141143412947153, "grad_norm": 0.42713966965675354, "learning_rate": 8e-05, "loss": 1.6838, "step": 7533 }, { "epoch": 0.8142224143521021, "grad_norm": 0.43264904618263245, "learning_rate": 8e-05, "loss": 1.6786, "step": 7534 }, { "epoch": 0.8143304874094888, "grad_norm": 0.45448368787765503, "learning_rate": 8e-05, "loss": 1.692, "step": 7535 }, { "epoch": 0.8144385604668756, "grad_norm": 0.43145474791526794, "learning_rate": 8e-05, "loss": 1.7434, "step": 7536 }, { "epoch": 0.8145466335242624, "grad_norm": 0.45540499687194824, "learning_rate": 8e-05, "loss": 1.6458, "step": 7537 }, { "epoch": 0.8146547065816492, "grad_norm": 0.4553711712360382, "learning_rate": 8e-05, "loss": 1.7855, "step": 7538 }, { "epoch": 0.814762779639036, "grad_norm": 0.45116326212882996, "learning_rate": 8e-05, "loss": 1.7935, "step": 7539 }, { "epoch": 0.8148708526964228, "grad_norm": 0.45032909512519836, "learning_rate": 8e-05, "loss": 1.6795, "step": 7540 }, { "epoch": 0.8149789257538096, "grad_norm": 0.4316040575504303, "learning_rate": 8e-05, "loss": 1.7853, "step": 7541 }, { "epoch": 0.8150869988111964, "grad_norm": 0.4131704568862915, "learning_rate": 8e-05, "loss": 1.4709, "step": 7542 }, { "epoch": 0.8151950718685832, "grad_norm": 0.46447908878326416, "learning_rate": 8e-05, "loss": 1.7019, "step": 7543 }, { "epoch": 0.81530314492597, "grad_norm": 0.415502667427063, "learning_rate": 8e-05, "loss": 1.654, "step": 7544 }, { "epoch": 0.8154112179833567, "grad_norm": 0.40690937638282776, "learning_rate": 8e-05, "loss": 1.6898, "step": 7545 }, { "epoch": 0.8155192910407435, "grad_norm": 0.4650536775588989, "learning_rate": 8e-05, "loss": 1.7662, "step": 7546 }, { "epoch": 0.8156273640981303, "grad_norm": 0.4277820885181427, "learning_rate": 8e-05, "loss": 1.626, "step": 7547 }, { "epoch": 0.8157354371555171, "grad_norm": 0.4531188905239105, "learning_rate": 8e-05, "loss": 1.5642, "step": 7548 }, { "epoch": 0.8158435102129039, "grad_norm": 0.4027608036994934, "learning_rate": 8e-05, "loss": 1.6757, "step": 7549 }, { "epoch": 0.8159515832702907, "grad_norm": 0.4521540701389313, "learning_rate": 8e-05, "loss": 1.6806, "step": 7550 }, { "epoch": 0.8160596563276775, "grad_norm": 0.48544925451278687, "learning_rate": 8e-05, "loss": 1.7373, "step": 7551 }, { "epoch": 0.8161677293850643, "grad_norm": 0.42520877718925476, "learning_rate": 8e-05, "loss": 1.6809, "step": 7552 }, { "epoch": 0.8162758024424511, "grad_norm": 0.4417572021484375, "learning_rate": 8e-05, "loss": 1.7163, "step": 7553 }, { "epoch": 0.8163838754998379, "grad_norm": 0.4538433253765106, "learning_rate": 8e-05, "loss": 1.5947, "step": 7554 }, { "epoch": 0.8164919485572247, "grad_norm": 0.4291155934333801, "learning_rate": 8e-05, "loss": 1.6787, "step": 7555 }, { "epoch": 0.8166000216146114, "grad_norm": 0.43404850363731384, "learning_rate": 8e-05, "loss": 1.6873, "step": 7556 }, { "epoch": 0.8167080946719982, "grad_norm": 0.4470057487487793, "learning_rate": 8e-05, "loss": 1.7592, "step": 7557 }, { "epoch": 0.816816167729385, "grad_norm": 0.46990513801574707, "learning_rate": 8e-05, "loss": 1.8532, "step": 7558 }, { "epoch": 0.8169242407867718, "grad_norm": 0.44352492690086365, "learning_rate": 8e-05, "loss": 1.6379, "step": 7559 }, { "epoch": 0.8170323138441586, "grad_norm": 0.4366074204444885, "learning_rate": 8e-05, "loss": 1.6068, "step": 7560 }, { "epoch": 0.8171403869015454, "grad_norm": 0.44509410858154297, "learning_rate": 8e-05, "loss": 1.7065, "step": 7561 }, { "epoch": 0.8172484599589322, "grad_norm": 0.45479580760002136, "learning_rate": 8e-05, "loss": 1.7363, "step": 7562 }, { "epoch": 0.817356533016319, "grad_norm": 0.4371161460876465, "learning_rate": 8e-05, "loss": 1.7584, "step": 7563 }, { "epoch": 0.8174646060737059, "grad_norm": 0.4582262337207794, "learning_rate": 8e-05, "loss": 1.7377, "step": 7564 }, { "epoch": 0.8175726791310927, "grad_norm": 0.40442079305648804, "learning_rate": 8e-05, "loss": 1.7332, "step": 7565 }, { "epoch": 0.8176807521884795, "grad_norm": 0.4005148112773895, "learning_rate": 8e-05, "loss": 1.5652, "step": 7566 }, { "epoch": 0.8177888252458662, "grad_norm": 0.4014653265476227, "learning_rate": 8e-05, "loss": 1.6576, "step": 7567 }, { "epoch": 0.817896898303253, "grad_norm": 0.5215381979942322, "learning_rate": 8e-05, "loss": 1.7231, "step": 7568 }, { "epoch": 0.8180049713606398, "grad_norm": 0.45727983117103577, "learning_rate": 8e-05, "loss": 1.5301, "step": 7569 }, { "epoch": 0.8181130444180266, "grad_norm": 0.42418810725212097, "learning_rate": 8e-05, "loss": 1.5554, "step": 7570 }, { "epoch": 0.8182211174754134, "grad_norm": 0.4791761338710785, "learning_rate": 8e-05, "loss": 1.7908, "step": 7571 }, { "epoch": 0.8183291905328002, "grad_norm": 0.4208248555660248, "learning_rate": 8e-05, "loss": 1.6958, "step": 7572 }, { "epoch": 0.818437263590187, "grad_norm": 0.43180933594703674, "learning_rate": 8e-05, "loss": 1.6126, "step": 7573 }, { "epoch": 0.8185453366475738, "grad_norm": 0.5061798095703125, "learning_rate": 8e-05, "loss": 1.5501, "step": 7574 }, { "epoch": 0.8186534097049606, "grad_norm": 0.42237475514411926, "learning_rate": 8e-05, "loss": 1.65, "step": 7575 }, { "epoch": 0.8187614827623474, "grad_norm": 0.45844408869743347, "learning_rate": 8e-05, "loss": 1.5858, "step": 7576 }, { "epoch": 0.8188695558197341, "grad_norm": 0.4064740240573883, "learning_rate": 8e-05, "loss": 1.5389, "step": 7577 }, { "epoch": 0.8189776288771209, "grad_norm": 0.4545830190181732, "learning_rate": 8e-05, "loss": 1.5942, "step": 7578 }, { "epoch": 0.8190857019345077, "grad_norm": 0.4766400456428528, "learning_rate": 8e-05, "loss": 1.7088, "step": 7579 }, { "epoch": 0.8191937749918945, "grad_norm": 0.4936043620109558, "learning_rate": 8e-05, "loss": 1.7938, "step": 7580 }, { "epoch": 0.8193018480492813, "grad_norm": 0.422983855009079, "learning_rate": 8e-05, "loss": 1.6012, "step": 7581 }, { "epoch": 0.8194099211066681, "grad_norm": 0.43010538816452026, "learning_rate": 8e-05, "loss": 1.4655, "step": 7582 }, { "epoch": 0.8195179941640549, "grad_norm": 0.40586230158805847, "learning_rate": 8e-05, "loss": 1.5983, "step": 7583 }, { "epoch": 0.8196260672214417, "grad_norm": 0.4420272409915924, "learning_rate": 8e-05, "loss": 1.6891, "step": 7584 }, { "epoch": 0.8197341402788285, "grad_norm": 0.4256305992603302, "learning_rate": 8e-05, "loss": 1.7482, "step": 7585 }, { "epoch": 0.8198422133362153, "grad_norm": 0.4285375773906708, "learning_rate": 8e-05, "loss": 1.6339, "step": 7586 }, { "epoch": 0.8199502863936021, "grad_norm": 0.4269183278083801, "learning_rate": 8e-05, "loss": 1.6648, "step": 7587 }, { "epoch": 0.8200583594509888, "grad_norm": 0.4068998098373413, "learning_rate": 8e-05, "loss": 1.5795, "step": 7588 }, { "epoch": 0.8201664325083756, "grad_norm": 0.47086605429649353, "learning_rate": 8e-05, "loss": 1.7805, "step": 7589 }, { "epoch": 0.8202745055657624, "grad_norm": 0.4224134087562561, "learning_rate": 8e-05, "loss": 1.6201, "step": 7590 }, { "epoch": 0.8203825786231492, "grad_norm": 0.585591197013855, "learning_rate": 8e-05, "loss": 1.6042, "step": 7591 }, { "epoch": 0.820490651680536, "grad_norm": 0.47477656602859497, "learning_rate": 8e-05, "loss": 1.6661, "step": 7592 }, { "epoch": 0.8205987247379228, "grad_norm": 0.4659790098667145, "learning_rate": 8e-05, "loss": 1.5738, "step": 7593 }, { "epoch": 0.8207067977953096, "grad_norm": 0.4542687237262726, "learning_rate": 8e-05, "loss": 1.5599, "step": 7594 }, { "epoch": 0.8208148708526964, "grad_norm": 0.4578531086444855, "learning_rate": 8e-05, "loss": 1.7093, "step": 7595 }, { "epoch": 0.8209229439100832, "grad_norm": 0.4282548725605011, "learning_rate": 8e-05, "loss": 1.5634, "step": 7596 }, { "epoch": 0.82103101696747, "grad_norm": 0.4077598452568054, "learning_rate": 8e-05, "loss": 1.5062, "step": 7597 }, { "epoch": 0.8211390900248569, "grad_norm": 0.402471661567688, "learning_rate": 8e-05, "loss": 1.4754, "step": 7598 }, { "epoch": 0.8212471630822435, "grad_norm": 0.45625683665275574, "learning_rate": 8e-05, "loss": 1.7793, "step": 7599 }, { "epoch": 0.8213552361396304, "grad_norm": 0.42153432965278625, "learning_rate": 8e-05, "loss": 1.611, "step": 7600 }, { "epoch": 0.8214633091970172, "grad_norm": 0.43028196692466736, "learning_rate": 8e-05, "loss": 1.6629, "step": 7601 }, { "epoch": 0.821571382254404, "grad_norm": 0.4275728166103363, "learning_rate": 8e-05, "loss": 1.5953, "step": 7602 }, { "epoch": 0.8216794553117908, "grad_norm": 0.40624111890792847, "learning_rate": 8e-05, "loss": 1.3889, "step": 7603 }, { "epoch": 0.8217875283691776, "grad_norm": 0.44380101561546326, "learning_rate": 8e-05, "loss": 1.7619, "step": 7604 }, { "epoch": 0.8218956014265644, "grad_norm": 0.42274564504623413, "learning_rate": 8e-05, "loss": 1.5697, "step": 7605 }, { "epoch": 0.8220036744839512, "grad_norm": 0.42080071568489075, "learning_rate": 8e-05, "loss": 1.4965, "step": 7606 }, { "epoch": 0.822111747541338, "grad_norm": 0.4152235686779022, "learning_rate": 8e-05, "loss": 1.6094, "step": 7607 }, { "epoch": 0.8222198205987248, "grad_norm": 0.45041483640670776, "learning_rate": 8e-05, "loss": 1.6466, "step": 7608 }, { "epoch": 0.8223278936561116, "grad_norm": 0.46574991941452026, "learning_rate": 8e-05, "loss": 1.6938, "step": 7609 }, { "epoch": 0.8224359667134983, "grad_norm": 0.4416710436344147, "learning_rate": 8e-05, "loss": 1.6514, "step": 7610 }, { "epoch": 0.8225440397708851, "grad_norm": 0.4607199430465698, "learning_rate": 8e-05, "loss": 1.7812, "step": 7611 }, { "epoch": 0.8226521128282719, "grad_norm": 0.46931275725364685, "learning_rate": 8e-05, "loss": 1.818, "step": 7612 }, { "epoch": 0.8227601858856587, "grad_norm": 0.40168964862823486, "learning_rate": 8e-05, "loss": 1.6631, "step": 7613 }, { "epoch": 0.8228682589430455, "grad_norm": 0.42526113986968994, "learning_rate": 8e-05, "loss": 1.6784, "step": 7614 }, { "epoch": 0.8229763320004323, "grad_norm": 0.5322404503822327, "learning_rate": 8e-05, "loss": 1.9883, "step": 7615 }, { "epoch": 0.8230844050578191, "grad_norm": 0.44564369320869446, "learning_rate": 8e-05, "loss": 1.564, "step": 7616 }, { "epoch": 0.8231924781152059, "grad_norm": 0.4304073452949524, "learning_rate": 8e-05, "loss": 1.6159, "step": 7617 }, { "epoch": 0.8233005511725927, "grad_norm": 0.4697522521018982, "learning_rate": 8e-05, "loss": 1.8845, "step": 7618 }, { "epoch": 0.8234086242299795, "grad_norm": 0.4925960302352905, "learning_rate": 8e-05, "loss": 1.8143, "step": 7619 }, { "epoch": 0.8235166972873662, "grad_norm": 0.4734884798526764, "learning_rate": 8e-05, "loss": 1.751, "step": 7620 }, { "epoch": 0.823624770344753, "grad_norm": 0.4896402955055237, "learning_rate": 8e-05, "loss": 1.6718, "step": 7621 }, { "epoch": 0.8237328434021398, "grad_norm": 0.48254162073135376, "learning_rate": 8e-05, "loss": 1.9896, "step": 7622 }, { "epoch": 0.8238409164595266, "grad_norm": 0.4698091447353363, "learning_rate": 8e-05, "loss": 1.8492, "step": 7623 }, { "epoch": 0.8239489895169134, "grad_norm": 0.45183265209198, "learning_rate": 8e-05, "loss": 1.6544, "step": 7624 }, { "epoch": 0.8240570625743002, "grad_norm": 0.46388962864875793, "learning_rate": 8e-05, "loss": 1.6508, "step": 7625 }, { "epoch": 0.824165135631687, "grad_norm": 0.4461500346660614, "learning_rate": 8e-05, "loss": 1.6318, "step": 7626 }, { "epoch": 0.8242732086890738, "grad_norm": 0.44111502170562744, "learning_rate": 8e-05, "loss": 1.6783, "step": 7627 }, { "epoch": 0.8243812817464606, "grad_norm": 0.4211145043373108, "learning_rate": 8e-05, "loss": 1.5788, "step": 7628 }, { "epoch": 0.8244893548038474, "grad_norm": 0.4259558916091919, "learning_rate": 8e-05, "loss": 1.494, "step": 7629 }, { "epoch": 0.8245974278612342, "grad_norm": 0.4389492869377136, "learning_rate": 8e-05, "loss": 1.675, "step": 7630 }, { "epoch": 0.8247055009186209, "grad_norm": 0.451225221157074, "learning_rate": 8e-05, "loss": 1.6403, "step": 7631 }, { "epoch": 0.8248135739760077, "grad_norm": 0.46733182668685913, "learning_rate": 8e-05, "loss": 1.7181, "step": 7632 }, { "epoch": 0.8249216470333945, "grad_norm": 0.5023539066314697, "learning_rate": 8e-05, "loss": 1.5784, "step": 7633 }, { "epoch": 0.8250297200907813, "grad_norm": 0.40966740250587463, "learning_rate": 8e-05, "loss": 1.444, "step": 7634 }, { "epoch": 0.8251377931481682, "grad_norm": 0.4408402442932129, "learning_rate": 8e-05, "loss": 1.6626, "step": 7635 }, { "epoch": 0.825245866205555, "grad_norm": 0.43984270095825195, "learning_rate": 8e-05, "loss": 1.8169, "step": 7636 }, { "epoch": 0.8253539392629418, "grad_norm": 0.4835735261440277, "learning_rate": 8e-05, "loss": 1.7807, "step": 7637 }, { "epoch": 0.8254620123203286, "grad_norm": 0.4143886864185333, "learning_rate": 8e-05, "loss": 1.5459, "step": 7638 }, { "epoch": 0.8255700853777154, "grad_norm": 0.5253183841705322, "learning_rate": 8e-05, "loss": 1.8696, "step": 7639 }, { "epoch": 0.8256781584351022, "grad_norm": 0.43455126881599426, "learning_rate": 8e-05, "loss": 1.6783, "step": 7640 }, { "epoch": 0.825786231492489, "grad_norm": 0.4542931616306305, "learning_rate": 8e-05, "loss": 1.5294, "step": 7641 }, { "epoch": 0.8258943045498757, "grad_norm": 0.42774462699890137, "learning_rate": 8e-05, "loss": 1.6217, "step": 7642 }, { "epoch": 0.8260023776072625, "grad_norm": 0.4802873432636261, "learning_rate": 8e-05, "loss": 1.6462, "step": 7643 }, { "epoch": 0.8261104506646493, "grad_norm": 0.42625316977500916, "learning_rate": 8e-05, "loss": 1.492, "step": 7644 }, { "epoch": 0.8262185237220361, "grad_norm": 0.4295862913131714, "learning_rate": 8e-05, "loss": 1.7514, "step": 7645 }, { "epoch": 0.8263265967794229, "grad_norm": 0.4813641905784607, "learning_rate": 8e-05, "loss": 1.8041, "step": 7646 }, { "epoch": 0.8264346698368097, "grad_norm": 0.43345722556114197, "learning_rate": 8e-05, "loss": 1.6337, "step": 7647 }, { "epoch": 0.8265427428941965, "grad_norm": 0.4188258945941925, "learning_rate": 8e-05, "loss": 1.6675, "step": 7648 }, { "epoch": 0.8266508159515833, "grad_norm": 0.4513920843601227, "learning_rate": 8e-05, "loss": 1.6333, "step": 7649 }, { "epoch": 0.8267588890089701, "grad_norm": 0.47560781240463257, "learning_rate": 8e-05, "loss": 1.6588, "step": 7650 }, { "epoch": 0.8268669620663569, "grad_norm": 0.4722648561000824, "learning_rate": 8e-05, "loss": 1.64, "step": 7651 }, { "epoch": 0.8269750351237436, "grad_norm": 0.45728784799575806, "learning_rate": 8e-05, "loss": 1.6912, "step": 7652 }, { "epoch": 0.8270831081811304, "grad_norm": 0.45039939880371094, "learning_rate": 8e-05, "loss": 1.7254, "step": 7653 }, { "epoch": 0.8271911812385172, "grad_norm": 0.46544143557548523, "learning_rate": 8e-05, "loss": 1.7667, "step": 7654 }, { "epoch": 0.827299254295904, "grad_norm": 0.4608057737350464, "learning_rate": 8e-05, "loss": 1.6874, "step": 7655 }, { "epoch": 0.8274073273532908, "grad_norm": 0.4532715082168579, "learning_rate": 8e-05, "loss": 1.5849, "step": 7656 }, { "epoch": 0.8275154004106776, "grad_norm": 0.44655391573905945, "learning_rate": 8e-05, "loss": 1.5961, "step": 7657 }, { "epoch": 0.8276234734680644, "grad_norm": 0.47374507784843445, "learning_rate": 8e-05, "loss": 1.8575, "step": 7658 }, { "epoch": 0.8277315465254512, "grad_norm": 0.4443801939487457, "learning_rate": 8e-05, "loss": 1.6599, "step": 7659 }, { "epoch": 0.827839619582838, "grad_norm": 0.44458162784576416, "learning_rate": 8e-05, "loss": 1.8354, "step": 7660 }, { "epoch": 0.8279476926402248, "grad_norm": 0.48529109358787537, "learning_rate": 8e-05, "loss": 1.568, "step": 7661 }, { "epoch": 0.8280557656976116, "grad_norm": 0.47455430030822754, "learning_rate": 8e-05, "loss": 1.81, "step": 7662 }, { "epoch": 0.8281638387549983, "grad_norm": 0.43204164505004883, "learning_rate": 8e-05, "loss": 1.6573, "step": 7663 }, { "epoch": 0.8282719118123851, "grad_norm": 0.41443923115730286, "learning_rate": 8e-05, "loss": 1.6604, "step": 7664 }, { "epoch": 0.8283799848697719, "grad_norm": 0.5125259757041931, "learning_rate": 8e-05, "loss": 1.8525, "step": 7665 }, { "epoch": 0.8284880579271587, "grad_norm": 0.4526307284832001, "learning_rate": 8e-05, "loss": 1.678, "step": 7666 }, { "epoch": 0.8285961309845455, "grad_norm": 0.4572529196739197, "learning_rate": 8e-05, "loss": 1.7402, "step": 7667 }, { "epoch": 0.8287042040419323, "grad_norm": 0.43805745244026184, "learning_rate": 8e-05, "loss": 1.7234, "step": 7668 }, { "epoch": 0.8288122770993192, "grad_norm": 0.44070878624916077, "learning_rate": 8e-05, "loss": 1.6876, "step": 7669 }, { "epoch": 0.828920350156706, "grad_norm": 0.40329840779304504, "learning_rate": 8e-05, "loss": 1.6791, "step": 7670 }, { "epoch": 0.8290284232140928, "grad_norm": 0.4230155050754547, "learning_rate": 8e-05, "loss": 1.6781, "step": 7671 }, { "epoch": 0.8291364962714796, "grad_norm": 0.40365561842918396, "learning_rate": 8e-05, "loss": 1.5906, "step": 7672 }, { "epoch": 0.8292445693288664, "grad_norm": 0.4522151052951813, "learning_rate": 8e-05, "loss": 1.8112, "step": 7673 }, { "epoch": 0.8293526423862531, "grad_norm": 0.45651283860206604, "learning_rate": 8e-05, "loss": 1.669, "step": 7674 }, { "epoch": 0.8294607154436399, "grad_norm": 0.44629165530204773, "learning_rate": 8e-05, "loss": 1.6349, "step": 7675 }, { "epoch": 0.8295687885010267, "grad_norm": 0.46019428968429565, "learning_rate": 8e-05, "loss": 1.8082, "step": 7676 }, { "epoch": 0.8296768615584135, "grad_norm": 0.3903871178627014, "learning_rate": 8e-05, "loss": 1.525, "step": 7677 }, { "epoch": 0.8297849346158003, "grad_norm": 0.4500637948513031, "learning_rate": 8e-05, "loss": 1.7378, "step": 7678 }, { "epoch": 0.8298930076731871, "grad_norm": 0.4200245440006256, "learning_rate": 8e-05, "loss": 1.5854, "step": 7679 }, { "epoch": 0.8300010807305739, "grad_norm": 0.4182475507259369, "learning_rate": 8e-05, "loss": 1.6378, "step": 7680 }, { "epoch": 0.8301091537879607, "grad_norm": 0.4354560077190399, "learning_rate": 8e-05, "loss": 1.6605, "step": 7681 }, { "epoch": 0.8302172268453475, "grad_norm": 0.5051718950271606, "learning_rate": 8e-05, "loss": 1.6287, "step": 7682 }, { "epoch": 0.8303252999027343, "grad_norm": 0.4145927131175995, "learning_rate": 8e-05, "loss": 1.664, "step": 7683 }, { "epoch": 0.8304333729601211, "grad_norm": 0.41116395592689514, "learning_rate": 8e-05, "loss": 1.5686, "step": 7684 }, { "epoch": 0.8305414460175078, "grad_norm": 0.44805383682250977, "learning_rate": 8e-05, "loss": 1.7488, "step": 7685 }, { "epoch": 0.8306495190748946, "grad_norm": 0.4370586574077606, "learning_rate": 8e-05, "loss": 1.5686, "step": 7686 }, { "epoch": 0.8307575921322814, "grad_norm": 0.4257923662662506, "learning_rate": 8e-05, "loss": 1.5673, "step": 7687 }, { "epoch": 0.8308656651896682, "grad_norm": 0.431093692779541, "learning_rate": 8e-05, "loss": 1.7079, "step": 7688 }, { "epoch": 0.830973738247055, "grad_norm": 0.44439396262168884, "learning_rate": 8e-05, "loss": 1.576, "step": 7689 }, { "epoch": 0.8310818113044418, "grad_norm": 0.45661646127700806, "learning_rate": 8e-05, "loss": 1.7491, "step": 7690 }, { "epoch": 0.8311898843618286, "grad_norm": 0.4075823426246643, "learning_rate": 8e-05, "loss": 1.6141, "step": 7691 }, { "epoch": 0.8312979574192154, "grad_norm": 0.41610363125801086, "learning_rate": 8e-05, "loss": 1.6587, "step": 7692 }, { "epoch": 0.8314060304766022, "grad_norm": 0.410784512758255, "learning_rate": 8e-05, "loss": 1.6744, "step": 7693 }, { "epoch": 0.831514103533989, "grad_norm": 0.46869975328445435, "learning_rate": 8e-05, "loss": 1.7702, "step": 7694 }, { "epoch": 0.8316221765913757, "grad_norm": 0.4682677984237671, "learning_rate": 8e-05, "loss": 1.5637, "step": 7695 }, { "epoch": 0.8317302496487625, "grad_norm": 0.41629067063331604, "learning_rate": 8e-05, "loss": 1.4837, "step": 7696 }, { "epoch": 0.8318383227061493, "grad_norm": 0.45000967383384705, "learning_rate": 8e-05, "loss": 1.6576, "step": 7697 }, { "epoch": 0.8319463957635361, "grad_norm": 0.4109426438808441, "learning_rate": 8e-05, "loss": 1.6012, "step": 7698 }, { "epoch": 0.8320544688209229, "grad_norm": 0.44350165128707886, "learning_rate": 8e-05, "loss": 1.6987, "step": 7699 }, { "epoch": 0.8321625418783097, "grad_norm": 0.42586860060691833, "learning_rate": 8e-05, "loss": 1.7608, "step": 7700 }, { "epoch": 0.8322706149356965, "grad_norm": 0.4208213686943054, "learning_rate": 8e-05, "loss": 1.8316, "step": 7701 }, { "epoch": 0.8323786879930833, "grad_norm": 0.39631909132003784, "learning_rate": 8e-05, "loss": 1.6782, "step": 7702 }, { "epoch": 0.8324867610504701, "grad_norm": 0.48009052872657776, "learning_rate": 8e-05, "loss": 1.8464, "step": 7703 }, { "epoch": 0.832594834107857, "grad_norm": 0.4532031714916229, "learning_rate": 8e-05, "loss": 1.8129, "step": 7704 }, { "epoch": 0.8327029071652438, "grad_norm": 0.43256875872612, "learning_rate": 8e-05, "loss": 1.6251, "step": 7705 }, { "epoch": 0.8328109802226304, "grad_norm": 0.4651573896408081, "learning_rate": 8e-05, "loss": 1.8131, "step": 7706 }, { "epoch": 0.8329190532800173, "grad_norm": 0.4159470498561859, "learning_rate": 8e-05, "loss": 1.5457, "step": 7707 }, { "epoch": 0.8330271263374041, "grad_norm": 0.4307110011577606, "learning_rate": 8e-05, "loss": 1.6888, "step": 7708 }, { "epoch": 0.8331351993947909, "grad_norm": 0.41106343269348145, "learning_rate": 8e-05, "loss": 1.7072, "step": 7709 }, { "epoch": 0.8332432724521777, "grad_norm": 0.42679154872894287, "learning_rate": 8e-05, "loss": 1.5411, "step": 7710 }, { "epoch": 0.8333513455095645, "grad_norm": 0.46125566959381104, "learning_rate": 8e-05, "loss": 1.6292, "step": 7711 }, { "epoch": 0.8334594185669513, "grad_norm": 0.44966283440589905, "learning_rate": 8e-05, "loss": 1.7467, "step": 7712 }, { "epoch": 0.8335674916243381, "grad_norm": 0.44620373845100403, "learning_rate": 8e-05, "loss": 1.678, "step": 7713 }, { "epoch": 0.8336755646817249, "grad_norm": 0.4552816152572632, "learning_rate": 8e-05, "loss": 1.6153, "step": 7714 }, { "epoch": 0.8337836377391117, "grad_norm": 0.5180802345275879, "learning_rate": 8e-05, "loss": 1.9585, "step": 7715 }, { "epoch": 0.8338917107964985, "grad_norm": 0.44563549757003784, "learning_rate": 8e-05, "loss": 1.6615, "step": 7716 }, { "epoch": 0.8339997838538852, "grad_norm": 0.4206632375717163, "learning_rate": 8e-05, "loss": 1.6477, "step": 7717 }, { "epoch": 0.834107856911272, "grad_norm": 0.4389001429080963, "learning_rate": 8e-05, "loss": 1.7486, "step": 7718 }, { "epoch": 0.8342159299686588, "grad_norm": 0.40575510263442993, "learning_rate": 8e-05, "loss": 1.5633, "step": 7719 }, { "epoch": 0.8343240030260456, "grad_norm": 0.43390125036239624, "learning_rate": 8e-05, "loss": 1.5458, "step": 7720 }, { "epoch": 0.8344320760834324, "grad_norm": 0.4245692789554596, "learning_rate": 8e-05, "loss": 1.5286, "step": 7721 }, { "epoch": 0.8345401491408192, "grad_norm": 0.44516414403915405, "learning_rate": 8e-05, "loss": 1.6026, "step": 7722 }, { "epoch": 0.834648222198206, "grad_norm": 0.4553352892398834, "learning_rate": 8e-05, "loss": 1.7343, "step": 7723 }, { "epoch": 0.8347562952555928, "grad_norm": 0.44706717133522034, "learning_rate": 8e-05, "loss": 1.8112, "step": 7724 }, { "epoch": 0.8348643683129796, "grad_norm": 0.46013206243515015, "learning_rate": 8e-05, "loss": 1.7568, "step": 7725 }, { "epoch": 0.8349724413703664, "grad_norm": 0.4639611542224884, "learning_rate": 8e-05, "loss": 1.8401, "step": 7726 }, { "epoch": 0.8350805144277531, "grad_norm": 0.45895493030548096, "learning_rate": 8e-05, "loss": 1.7113, "step": 7727 }, { "epoch": 0.8351885874851399, "grad_norm": 0.4412960112094879, "learning_rate": 8e-05, "loss": 1.7817, "step": 7728 }, { "epoch": 0.8352966605425267, "grad_norm": 0.4324769079685211, "learning_rate": 8e-05, "loss": 1.4806, "step": 7729 }, { "epoch": 0.8354047335999135, "grad_norm": 0.4650159776210785, "learning_rate": 8e-05, "loss": 1.7733, "step": 7730 }, { "epoch": 0.8355128066573003, "grad_norm": 0.41595304012298584, "learning_rate": 8e-05, "loss": 1.621, "step": 7731 }, { "epoch": 0.8356208797146871, "grad_norm": 0.448150098323822, "learning_rate": 8e-05, "loss": 1.59, "step": 7732 }, { "epoch": 0.8357289527720739, "grad_norm": 0.4187316298484802, "learning_rate": 8e-05, "loss": 1.6002, "step": 7733 }, { "epoch": 0.8358370258294607, "grad_norm": 0.43724653124809265, "learning_rate": 8e-05, "loss": 1.6039, "step": 7734 }, { "epoch": 0.8359450988868475, "grad_norm": 0.44794124364852905, "learning_rate": 8e-05, "loss": 1.7363, "step": 7735 }, { "epoch": 0.8360531719442343, "grad_norm": 0.4605340361595154, "learning_rate": 8e-05, "loss": 1.6693, "step": 7736 }, { "epoch": 0.8361612450016211, "grad_norm": 0.4455616772174835, "learning_rate": 8e-05, "loss": 1.7438, "step": 7737 }, { "epoch": 0.8362693180590078, "grad_norm": 0.46097931265830994, "learning_rate": 8e-05, "loss": 1.7587, "step": 7738 }, { "epoch": 0.8363773911163946, "grad_norm": 0.424600213766098, "learning_rate": 8e-05, "loss": 1.6491, "step": 7739 }, { "epoch": 0.8364854641737814, "grad_norm": 0.4342171847820282, "learning_rate": 8e-05, "loss": 1.6536, "step": 7740 }, { "epoch": 0.8365935372311682, "grad_norm": 0.4803837537765503, "learning_rate": 8e-05, "loss": 1.8358, "step": 7741 }, { "epoch": 0.836701610288555, "grad_norm": 0.44699573516845703, "learning_rate": 8e-05, "loss": 1.8694, "step": 7742 }, { "epoch": 0.8368096833459419, "grad_norm": 0.41732078790664673, "learning_rate": 8e-05, "loss": 1.579, "step": 7743 }, { "epoch": 0.8369177564033287, "grad_norm": 0.4626522362232208, "learning_rate": 8e-05, "loss": 1.7366, "step": 7744 }, { "epoch": 0.8370258294607155, "grad_norm": 0.418046236038208, "learning_rate": 8e-05, "loss": 1.5956, "step": 7745 }, { "epoch": 0.8371339025181023, "grad_norm": 0.48796984553337097, "learning_rate": 8e-05, "loss": 1.7282, "step": 7746 }, { "epoch": 0.8372419755754891, "grad_norm": 0.47932907938957214, "learning_rate": 8e-05, "loss": 1.8394, "step": 7747 }, { "epoch": 0.8373500486328759, "grad_norm": 0.4287440776824951, "learning_rate": 8e-05, "loss": 1.6656, "step": 7748 }, { "epoch": 0.8374581216902626, "grad_norm": 0.43780845403671265, "learning_rate": 8e-05, "loss": 1.6507, "step": 7749 }, { "epoch": 0.8375661947476494, "grad_norm": 0.43937695026397705, "learning_rate": 8e-05, "loss": 1.629, "step": 7750 }, { "epoch": 0.8376742678050362, "grad_norm": 0.4372697174549103, "learning_rate": 8e-05, "loss": 1.6178, "step": 7751 }, { "epoch": 0.837782340862423, "grad_norm": 0.43558230996131897, "learning_rate": 8e-05, "loss": 1.5758, "step": 7752 }, { "epoch": 0.8378904139198098, "grad_norm": 0.43865111470222473, "learning_rate": 8e-05, "loss": 1.7966, "step": 7753 }, { "epoch": 0.8379984869771966, "grad_norm": 0.449001669883728, "learning_rate": 8e-05, "loss": 1.7072, "step": 7754 }, { "epoch": 0.8381065600345834, "grad_norm": 0.4586077928543091, "learning_rate": 8e-05, "loss": 1.7107, "step": 7755 }, { "epoch": 0.8382146330919702, "grad_norm": 0.43268027901649475, "learning_rate": 8e-05, "loss": 1.6721, "step": 7756 }, { "epoch": 0.838322706149357, "grad_norm": 0.44699281454086304, "learning_rate": 8e-05, "loss": 1.6231, "step": 7757 }, { "epoch": 0.8384307792067438, "grad_norm": 0.4412274658679962, "learning_rate": 8e-05, "loss": 1.6922, "step": 7758 }, { "epoch": 0.8385388522641306, "grad_norm": 0.42755597829818726, "learning_rate": 8e-05, "loss": 1.6171, "step": 7759 }, { "epoch": 0.8386469253215173, "grad_norm": 0.46667370200157166, "learning_rate": 8e-05, "loss": 1.7705, "step": 7760 }, { "epoch": 0.8387549983789041, "grad_norm": 0.44554901123046875, "learning_rate": 8e-05, "loss": 1.7475, "step": 7761 }, { "epoch": 0.8388630714362909, "grad_norm": 0.4607847332954407, "learning_rate": 8e-05, "loss": 1.6481, "step": 7762 }, { "epoch": 0.8389711444936777, "grad_norm": 0.46327337622642517, "learning_rate": 8e-05, "loss": 1.6033, "step": 7763 }, { "epoch": 0.8390792175510645, "grad_norm": 0.4606308043003082, "learning_rate": 8e-05, "loss": 1.7185, "step": 7764 }, { "epoch": 0.8391872906084513, "grad_norm": 0.4502409100532532, "learning_rate": 8e-05, "loss": 1.7612, "step": 7765 }, { "epoch": 0.8392953636658381, "grad_norm": 0.4446258842945099, "learning_rate": 8e-05, "loss": 1.613, "step": 7766 }, { "epoch": 0.8394034367232249, "grad_norm": 0.4867793023586273, "learning_rate": 8e-05, "loss": 1.6403, "step": 7767 }, { "epoch": 0.8395115097806117, "grad_norm": 0.45607608556747437, "learning_rate": 8e-05, "loss": 1.6547, "step": 7768 }, { "epoch": 0.8396195828379985, "grad_norm": 0.45544517040252686, "learning_rate": 8e-05, "loss": 1.8247, "step": 7769 }, { "epoch": 0.8397276558953852, "grad_norm": 0.4399893581867218, "learning_rate": 8e-05, "loss": 1.7513, "step": 7770 }, { "epoch": 0.839835728952772, "grad_norm": 0.43779149651527405, "learning_rate": 8e-05, "loss": 1.6182, "step": 7771 }, { "epoch": 0.8399438020101588, "grad_norm": 0.4638163447380066, "learning_rate": 8e-05, "loss": 1.8614, "step": 7772 }, { "epoch": 0.8400518750675456, "grad_norm": 0.4200887382030487, "learning_rate": 8e-05, "loss": 1.5931, "step": 7773 }, { "epoch": 0.8401599481249324, "grad_norm": 0.4699588716030121, "learning_rate": 8e-05, "loss": 1.7871, "step": 7774 }, { "epoch": 0.8402680211823192, "grad_norm": 0.4680885970592499, "learning_rate": 8e-05, "loss": 1.8299, "step": 7775 }, { "epoch": 0.840376094239706, "grad_norm": 0.4250168204307556, "learning_rate": 8e-05, "loss": 1.7256, "step": 7776 }, { "epoch": 0.8404841672970929, "grad_norm": 0.4655633568763733, "learning_rate": 8e-05, "loss": 1.6835, "step": 7777 }, { "epoch": 0.8405922403544797, "grad_norm": 0.44237154722213745, "learning_rate": 8e-05, "loss": 1.6869, "step": 7778 }, { "epoch": 0.8407003134118665, "grad_norm": 0.4061606526374817, "learning_rate": 8e-05, "loss": 1.6661, "step": 7779 }, { "epoch": 0.8408083864692533, "grad_norm": 0.42082127928733826, "learning_rate": 8e-05, "loss": 1.4393, "step": 7780 }, { "epoch": 0.84091645952664, "grad_norm": 0.4313361644744873, "learning_rate": 8e-05, "loss": 1.5411, "step": 7781 }, { "epoch": 0.8410245325840268, "grad_norm": 0.4462850093841553, "learning_rate": 8e-05, "loss": 1.7271, "step": 7782 }, { "epoch": 0.8411326056414136, "grad_norm": 0.4583030045032501, "learning_rate": 8e-05, "loss": 1.7734, "step": 7783 }, { "epoch": 0.8412406786988004, "grad_norm": 0.4669739603996277, "learning_rate": 8e-05, "loss": 1.5003, "step": 7784 }, { "epoch": 0.8413487517561872, "grad_norm": 0.4741883873939514, "learning_rate": 8e-05, "loss": 1.5634, "step": 7785 }, { "epoch": 0.841456824813574, "grad_norm": 0.4066810607910156, "learning_rate": 8e-05, "loss": 1.4923, "step": 7786 }, { "epoch": 0.8415648978709608, "grad_norm": 0.42306554317474365, "learning_rate": 8e-05, "loss": 1.6022, "step": 7787 }, { "epoch": 0.8416729709283476, "grad_norm": 0.42263513803482056, "learning_rate": 8e-05, "loss": 1.4847, "step": 7788 }, { "epoch": 0.8417810439857344, "grad_norm": 0.45076045393943787, "learning_rate": 8e-05, "loss": 1.6912, "step": 7789 }, { "epoch": 0.8418891170431212, "grad_norm": 0.4528537392616272, "learning_rate": 8e-05, "loss": 1.6806, "step": 7790 }, { "epoch": 0.841997190100508, "grad_norm": 0.4536552429199219, "learning_rate": 8e-05, "loss": 1.6093, "step": 7791 }, { "epoch": 0.8421052631578947, "grad_norm": 0.42940816283226013, "learning_rate": 8e-05, "loss": 1.6279, "step": 7792 }, { "epoch": 0.8422133362152815, "grad_norm": 0.44882044196128845, "learning_rate": 8e-05, "loss": 1.6931, "step": 7793 }, { "epoch": 0.8423214092726683, "grad_norm": 0.4309455454349518, "learning_rate": 8e-05, "loss": 1.6307, "step": 7794 }, { "epoch": 0.8424294823300551, "grad_norm": 0.4313630759716034, "learning_rate": 8e-05, "loss": 1.6748, "step": 7795 }, { "epoch": 0.8425375553874419, "grad_norm": 0.450801283121109, "learning_rate": 8e-05, "loss": 1.6702, "step": 7796 }, { "epoch": 0.8426456284448287, "grad_norm": 0.4354989528656006, "learning_rate": 8e-05, "loss": 1.5937, "step": 7797 }, { "epoch": 0.8427537015022155, "grad_norm": 0.4319075047969818, "learning_rate": 8e-05, "loss": 1.663, "step": 7798 }, { "epoch": 0.8428617745596023, "grad_norm": 0.43124303221702576, "learning_rate": 8e-05, "loss": 1.4283, "step": 7799 }, { "epoch": 0.8429698476169891, "grad_norm": 0.43562719225883484, "learning_rate": 8e-05, "loss": 1.7402, "step": 7800 }, { "epoch": 0.8430779206743759, "grad_norm": 0.45638564229011536, "learning_rate": 8e-05, "loss": 1.6212, "step": 7801 }, { "epoch": 0.8431859937317626, "grad_norm": 0.4562820494174957, "learning_rate": 8e-05, "loss": 1.6994, "step": 7802 }, { "epoch": 0.8432940667891494, "grad_norm": 0.4560452997684479, "learning_rate": 8e-05, "loss": 1.4878, "step": 7803 }, { "epoch": 0.8434021398465362, "grad_norm": 0.48244959115982056, "learning_rate": 8e-05, "loss": 1.7464, "step": 7804 }, { "epoch": 0.843510212903923, "grad_norm": 0.46198272705078125, "learning_rate": 8e-05, "loss": 1.7669, "step": 7805 }, { "epoch": 0.8436182859613098, "grad_norm": 0.42713865637779236, "learning_rate": 8e-05, "loss": 1.4386, "step": 7806 }, { "epoch": 0.8437263590186966, "grad_norm": 0.41599488258361816, "learning_rate": 8e-05, "loss": 1.5784, "step": 7807 }, { "epoch": 0.8438344320760834, "grad_norm": 0.4522525370121002, "learning_rate": 8e-05, "loss": 1.7763, "step": 7808 }, { "epoch": 0.8439425051334702, "grad_norm": 0.5152370929718018, "learning_rate": 8e-05, "loss": 1.9745, "step": 7809 }, { "epoch": 0.844050578190857, "grad_norm": 0.4332713782787323, "learning_rate": 8e-05, "loss": 1.4929, "step": 7810 }, { "epoch": 0.8441586512482439, "grad_norm": 0.46825891733169556, "learning_rate": 8e-05, "loss": 1.6858, "step": 7811 }, { "epoch": 0.8442667243056307, "grad_norm": 0.4517103135585785, "learning_rate": 8e-05, "loss": 1.6532, "step": 7812 }, { "epoch": 0.8443747973630173, "grad_norm": 0.4461933672428131, "learning_rate": 8e-05, "loss": 1.6409, "step": 7813 }, { "epoch": 0.8444828704204042, "grad_norm": 0.417714387178421, "learning_rate": 8e-05, "loss": 1.5822, "step": 7814 }, { "epoch": 0.844590943477791, "grad_norm": 0.5401750206947327, "learning_rate": 8e-05, "loss": 1.8568, "step": 7815 }, { "epoch": 0.8446990165351778, "grad_norm": 0.42826351523399353, "learning_rate": 8e-05, "loss": 1.6804, "step": 7816 }, { "epoch": 0.8448070895925646, "grad_norm": 0.42896512150764465, "learning_rate": 8e-05, "loss": 1.4911, "step": 7817 }, { "epoch": 0.8449151626499514, "grad_norm": 0.4434468746185303, "learning_rate": 8e-05, "loss": 1.756, "step": 7818 }, { "epoch": 0.8450232357073382, "grad_norm": 0.42869722843170166, "learning_rate": 8e-05, "loss": 1.723, "step": 7819 }, { "epoch": 0.845131308764725, "grad_norm": 0.442838191986084, "learning_rate": 8e-05, "loss": 1.7005, "step": 7820 }, { "epoch": 0.8452393818221118, "grad_norm": 0.4571494162082672, "learning_rate": 8e-05, "loss": 1.7715, "step": 7821 }, { "epoch": 0.8453474548794986, "grad_norm": 0.4711655378341675, "learning_rate": 8e-05, "loss": 1.5332, "step": 7822 }, { "epoch": 0.8454555279368854, "grad_norm": 0.4268406331539154, "learning_rate": 8e-05, "loss": 1.5528, "step": 7823 }, { "epoch": 0.8455636009942721, "grad_norm": 0.42949655652046204, "learning_rate": 8e-05, "loss": 1.76, "step": 7824 }, { "epoch": 0.8456716740516589, "grad_norm": 0.4312434494495392, "learning_rate": 8e-05, "loss": 1.5946, "step": 7825 }, { "epoch": 0.8457797471090457, "grad_norm": 0.42342042922973633, "learning_rate": 8e-05, "loss": 1.601, "step": 7826 }, { "epoch": 0.8458878201664325, "grad_norm": 0.4425911009311676, "learning_rate": 8e-05, "loss": 1.7761, "step": 7827 }, { "epoch": 0.8459958932238193, "grad_norm": 0.4829983115196228, "learning_rate": 8e-05, "loss": 1.8863, "step": 7828 }, { "epoch": 0.8461039662812061, "grad_norm": 0.4190134108066559, "learning_rate": 8e-05, "loss": 1.6605, "step": 7829 }, { "epoch": 0.8462120393385929, "grad_norm": 0.5107550024986267, "learning_rate": 8e-05, "loss": 1.7514, "step": 7830 }, { "epoch": 0.8463201123959797, "grad_norm": 0.45772451162338257, "learning_rate": 8e-05, "loss": 1.7055, "step": 7831 }, { "epoch": 0.8464281854533665, "grad_norm": 0.42696425318717957, "learning_rate": 8e-05, "loss": 1.7093, "step": 7832 }, { "epoch": 0.8465362585107533, "grad_norm": 0.40860220789909363, "learning_rate": 8e-05, "loss": 1.6913, "step": 7833 }, { "epoch": 0.84664433156814, "grad_norm": 0.4264879822731018, "learning_rate": 8e-05, "loss": 1.7642, "step": 7834 }, { "epoch": 0.8467524046255268, "grad_norm": 0.4342679977416992, "learning_rate": 8e-05, "loss": 1.6503, "step": 7835 }, { "epoch": 0.8468604776829136, "grad_norm": 0.41129592061042786, "learning_rate": 8e-05, "loss": 1.6714, "step": 7836 }, { "epoch": 0.8469685507403004, "grad_norm": 0.43578121066093445, "learning_rate": 8e-05, "loss": 1.6864, "step": 7837 }, { "epoch": 0.8470766237976872, "grad_norm": 0.42573654651641846, "learning_rate": 8e-05, "loss": 1.7367, "step": 7838 }, { "epoch": 0.847184696855074, "grad_norm": 0.42617014050483704, "learning_rate": 8e-05, "loss": 1.5867, "step": 7839 }, { "epoch": 0.8472927699124608, "grad_norm": 0.4175088405609131, "learning_rate": 8e-05, "loss": 1.7607, "step": 7840 }, { "epoch": 0.8474008429698476, "grad_norm": 0.44013676047325134, "learning_rate": 8e-05, "loss": 1.6843, "step": 7841 }, { "epoch": 0.8475089160272344, "grad_norm": 0.4078371226787567, "learning_rate": 8e-05, "loss": 1.5253, "step": 7842 }, { "epoch": 0.8476169890846212, "grad_norm": 0.43500739336013794, "learning_rate": 8e-05, "loss": 1.7076, "step": 7843 }, { "epoch": 0.847725062142008, "grad_norm": 0.3917448818683624, "learning_rate": 8e-05, "loss": 1.4377, "step": 7844 }, { "epoch": 0.8478331351993947, "grad_norm": 0.41684669256210327, "learning_rate": 8e-05, "loss": 1.5982, "step": 7845 }, { "epoch": 0.8479412082567815, "grad_norm": 0.46315279603004456, "learning_rate": 8e-05, "loss": 1.7701, "step": 7846 }, { "epoch": 0.8480492813141683, "grad_norm": 0.4571758210659027, "learning_rate": 8e-05, "loss": 1.6641, "step": 7847 }, { "epoch": 0.8481573543715551, "grad_norm": 0.4478438198566437, "learning_rate": 8e-05, "loss": 1.7313, "step": 7848 }, { "epoch": 0.848265427428942, "grad_norm": 0.43267467617988586, "learning_rate": 8e-05, "loss": 1.7569, "step": 7849 }, { "epoch": 0.8483735004863288, "grad_norm": 0.4610992670059204, "learning_rate": 8e-05, "loss": 1.5986, "step": 7850 }, { "epoch": 0.8484815735437156, "grad_norm": 0.450685977935791, "learning_rate": 8e-05, "loss": 1.6283, "step": 7851 }, { "epoch": 0.8485896466011024, "grad_norm": 0.45874688029289246, "learning_rate": 8e-05, "loss": 1.7813, "step": 7852 }, { "epoch": 0.8486977196584892, "grad_norm": 0.5470160841941833, "learning_rate": 8e-05, "loss": 2.1244, "step": 7853 }, { "epoch": 0.848805792715876, "grad_norm": 0.45659124851226807, "learning_rate": 8e-05, "loss": 1.6891, "step": 7854 }, { "epoch": 0.8489138657732628, "grad_norm": 0.40807920694351196, "learning_rate": 8e-05, "loss": 1.5308, "step": 7855 }, { "epoch": 0.8490219388306495, "grad_norm": 0.4221351742744446, "learning_rate": 8e-05, "loss": 1.5122, "step": 7856 }, { "epoch": 0.8491300118880363, "grad_norm": 0.4203934669494629, "learning_rate": 8e-05, "loss": 1.577, "step": 7857 }, { "epoch": 0.8492380849454231, "grad_norm": 0.47122636437416077, "learning_rate": 8e-05, "loss": 1.6904, "step": 7858 }, { "epoch": 0.8493461580028099, "grad_norm": 0.41796886920928955, "learning_rate": 8e-05, "loss": 1.5181, "step": 7859 }, { "epoch": 0.8494542310601967, "grad_norm": 0.41943585872650146, "learning_rate": 8e-05, "loss": 1.6923, "step": 7860 }, { "epoch": 0.8495623041175835, "grad_norm": 0.4859902262687683, "learning_rate": 8e-05, "loss": 1.7527, "step": 7861 }, { "epoch": 0.8496703771749703, "grad_norm": 0.460665225982666, "learning_rate": 8e-05, "loss": 1.6832, "step": 7862 }, { "epoch": 0.8497784502323571, "grad_norm": 0.47140470147132874, "learning_rate": 8e-05, "loss": 1.7982, "step": 7863 }, { "epoch": 0.8498865232897439, "grad_norm": 0.4585709273815155, "learning_rate": 8e-05, "loss": 1.627, "step": 7864 }, { "epoch": 0.8499945963471307, "grad_norm": 0.4501160979270935, "learning_rate": 8e-05, "loss": 1.7248, "step": 7865 }, { "epoch": 0.8501026694045175, "grad_norm": 0.4592576324939728, "learning_rate": 8e-05, "loss": 1.6656, "step": 7866 }, { "epoch": 0.8502107424619042, "grad_norm": 0.44173872470855713, "learning_rate": 8e-05, "loss": 1.5998, "step": 7867 }, { "epoch": 0.850318815519291, "grad_norm": 0.44532838463783264, "learning_rate": 8e-05, "loss": 1.5451, "step": 7868 }, { "epoch": 0.8504268885766778, "grad_norm": 0.42770448327064514, "learning_rate": 8e-05, "loss": 1.5045, "step": 7869 }, { "epoch": 0.8505349616340646, "grad_norm": 0.5034787654876709, "learning_rate": 8e-05, "loss": 1.7378, "step": 7870 }, { "epoch": 0.8506430346914514, "grad_norm": 0.45369991660118103, "learning_rate": 8e-05, "loss": 1.7631, "step": 7871 }, { "epoch": 0.8507511077488382, "grad_norm": 0.5151268839836121, "learning_rate": 8e-05, "loss": 1.8728, "step": 7872 }, { "epoch": 0.850859180806225, "grad_norm": 0.4396055042743683, "learning_rate": 8e-05, "loss": 1.6856, "step": 7873 }, { "epoch": 0.8509672538636118, "grad_norm": 0.45266807079315186, "learning_rate": 8e-05, "loss": 1.7951, "step": 7874 }, { "epoch": 0.8510753269209986, "grad_norm": 0.5143458843231201, "learning_rate": 8e-05, "loss": 1.7615, "step": 7875 }, { "epoch": 0.8511833999783854, "grad_norm": 0.4505205750465393, "learning_rate": 8e-05, "loss": 1.7001, "step": 7876 }, { "epoch": 0.8512914730357721, "grad_norm": 0.4212293326854706, "learning_rate": 8e-05, "loss": 1.4935, "step": 7877 }, { "epoch": 0.8513995460931589, "grad_norm": 0.42941275238990784, "learning_rate": 8e-05, "loss": 1.6759, "step": 7878 }, { "epoch": 0.8515076191505457, "grad_norm": 0.4792408347129822, "learning_rate": 8e-05, "loss": 1.9617, "step": 7879 }, { "epoch": 0.8516156922079325, "grad_norm": 0.4287477135658264, "learning_rate": 8e-05, "loss": 1.65, "step": 7880 }, { "epoch": 0.8517237652653193, "grad_norm": 0.45571771264076233, "learning_rate": 8e-05, "loss": 1.7236, "step": 7881 }, { "epoch": 0.8518318383227061, "grad_norm": 0.4141734838485718, "learning_rate": 8e-05, "loss": 1.6062, "step": 7882 }, { "epoch": 0.851939911380093, "grad_norm": 0.4653204083442688, "learning_rate": 8e-05, "loss": 1.5945, "step": 7883 }, { "epoch": 0.8520479844374798, "grad_norm": 0.4213482737541199, "learning_rate": 8e-05, "loss": 1.5971, "step": 7884 }, { "epoch": 0.8521560574948666, "grad_norm": 0.4579106271266937, "learning_rate": 8e-05, "loss": 1.6597, "step": 7885 }, { "epoch": 0.8522641305522534, "grad_norm": 0.46371474862098694, "learning_rate": 8e-05, "loss": 1.7255, "step": 7886 }, { "epoch": 0.8523722036096402, "grad_norm": 0.4525402784347534, "learning_rate": 8e-05, "loss": 1.7039, "step": 7887 }, { "epoch": 0.8524802766670269, "grad_norm": 0.6035670042037964, "learning_rate": 8e-05, "loss": 1.7863, "step": 7888 }, { "epoch": 0.8525883497244137, "grad_norm": 0.44751542806625366, "learning_rate": 8e-05, "loss": 1.5491, "step": 7889 }, { "epoch": 0.8526964227818005, "grad_norm": 0.5202107429504395, "learning_rate": 8e-05, "loss": 1.6712, "step": 7890 }, { "epoch": 0.8528044958391873, "grad_norm": 0.45437657833099365, "learning_rate": 8e-05, "loss": 1.6739, "step": 7891 }, { "epoch": 0.8529125688965741, "grad_norm": 0.45621949434280396, "learning_rate": 8e-05, "loss": 1.48, "step": 7892 }, { "epoch": 0.8530206419539609, "grad_norm": 0.5064789056777954, "learning_rate": 8e-05, "loss": 1.7961, "step": 7893 }, { "epoch": 0.8531287150113477, "grad_norm": 0.44472166895866394, "learning_rate": 8e-05, "loss": 1.682, "step": 7894 }, { "epoch": 0.8532367880687345, "grad_norm": 0.4199577867984772, "learning_rate": 8e-05, "loss": 1.5355, "step": 7895 }, { "epoch": 0.8533448611261213, "grad_norm": 0.41415467858314514, "learning_rate": 8e-05, "loss": 1.5549, "step": 7896 }, { "epoch": 0.8534529341835081, "grad_norm": 0.4585660696029663, "learning_rate": 8e-05, "loss": 1.6821, "step": 7897 }, { "epoch": 0.8535610072408949, "grad_norm": 0.4559144973754883, "learning_rate": 8e-05, "loss": 1.6734, "step": 7898 }, { "epoch": 0.8536690802982816, "grad_norm": 0.4823892414569855, "learning_rate": 8e-05, "loss": 1.7489, "step": 7899 }, { "epoch": 0.8537771533556684, "grad_norm": 0.4860594570636749, "learning_rate": 8e-05, "loss": 1.7057, "step": 7900 }, { "epoch": 0.8538852264130552, "grad_norm": 0.4262695908546448, "learning_rate": 8e-05, "loss": 1.6321, "step": 7901 }, { "epoch": 0.853993299470442, "grad_norm": 0.4338065981864929, "learning_rate": 8e-05, "loss": 1.736, "step": 7902 }, { "epoch": 0.8541013725278288, "grad_norm": 0.43874669075012207, "learning_rate": 8e-05, "loss": 1.5861, "step": 7903 }, { "epoch": 0.8542094455852156, "grad_norm": 0.48086661100387573, "learning_rate": 8e-05, "loss": 1.7992, "step": 7904 }, { "epoch": 0.8543175186426024, "grad_norm": 0.43347322940826416, "learning_rate": 8e-05, "loss": 1.5722, "step": 7905 }, { "epoch": 0.8544255916999892, "grad_norm": 0.42598825693130493, "learning_rate": 8e-05, "loss": 1.6058, "step": 7906 }, { "epoch": 0.854533664757376, "grad_norm": 0.4431690573692322, "learning_rate": 8e-05, "loss": 1.6476, "step": 7907 }, { "epoch": 0.8546417378147628, "grad_norm": 0.46456605195999146, "learning_rate": 8e-05, "loss": 1.8223, "step": 7908 }, { "epoch": 0.8547498108721495, "grad_norm": 0.45826172828674316, "learning_rate": 8e-05, "loss": 1.6247, "step": 7909 }, { "epoch": 0.8548578839295363, "grad_norm": 0.42410778999328613, "learning_rate": 8e-05, "loss": 1.6939, "step": 7910 }, { "epoch": 0.8549659569869231, "grad_norm": 0.422543466091156, "learning_rate": 8e-05, "loss": 1.5462, "step": 7911 }, { "epoch": 0.8550740300443099, "grad_norm": 0.47523176670074463, "learning_rate": 8e-05, "loss": 1.7677, "step": 7912 }, { "epoch": 0.8551821031016967, "grad_norm": 0.4378065764904022, "learning_rate": 8e-05, "loss": 1.6878, "step": 7913 }, { "epoch": 0.8552901761590835, "grad_norm": 0.4690306484699249, "learning_rate": 8e-05, "loss": 1.7536, "step": 7914 }, { "epoch": 0.8553982492164703, "grad_norm": 0.4486580491065979, "learning_rate": 8e-05, "loss": 1.5519, "step": 7915 }, { "epoch": 0.8555063222738571, "grad_norm": 0.5140929818153381, "learning_rate": 8e-05, "loss": 1.599, "step": 7916 }, { "epoch": 0.855614395331244, "grad_norm": 0.4226457178592682, "learning_rate": 8e-05, "loss": 1.5911, "step": 7917 }, { "epoch": 0.8557224683886308, "grad_norm": 0.4092429578304291, "learning_rate": 8e-05, "loss": 1.6199, "step": 7918 }, { "epoch": 0.8558305414460176, "grad_norm": 0.45524176955223083, "learning_rate": 8e-05, "loss": 1.7286, "step": 7919 }, { "epoch": 0.8559386145034042, "grad_norm": 0.4609297215938568, "learning_rate": 8e-05, "loss": 1.7685, "step": 7920 }, { "epoch": 0.856046687560791, "grad_norm": 0.4525575339794159, "learning_rate": 8e-05, "loss": 1.741, "step": 7921 }, { "epoch": 0.8561547606181779, "grad_norm": 0.4552176296710968, "learning_rate": 8e-05, "loss": 1.7074, "step": 7922 }, { "epoch": 0.8562628336755647, "grad_norm": 0.389146625995636, "learning_rate": 8e-05, "loss": 1.4691, "step": 7923 }, { "epoch": 0.8563709067329515, "grad_norm": 0.531312882900238, "learning_rate": 8e-05, "loss": 1.8525, "step": 7924 }, { "epoch": 0.8564789797903383, "grad_norm": 0.40663430094718933, "learning_rate": 8e-05, "loss": 1.635, "step": 7925 }, { "epoch": 0.8565870528477251, "grad_norm": 0.45749035477638245, "learning_rate": 8e-05, "loss": 1.6625, "step": 7926 }, { "epoch": 0.8566951259051119, "grad_norm": 0.44372883439064026, "learning_rate": 8e-05, "loss": 1.5991, "step": 7927 }, { "epoch": 0.8568031989624987, "grad_norm": 0.4802502691745758, "learning_rate": 8e-05, "loss": 1.766, "step": 7928 }, { "epoch": 0.8569112720198855, "grad_norm": 0.439615935087204, "learning_rate": 8e-05, "loss": 1.4654, "step": 7929 }, { "epoch": 0.8570193450772723, "grad_norm": 0.43785223364830017, "learning_rate": 8e-05, "loss": 1.6406, "step": 7930 }, { "epoch": 0.857127418134659, "grad_norm": 0.5155622363090515, "learning_rate": 8e-05, "loss": 1.6016, "step": 7931 }, { "epoch": 0.8572354911920458, "grad_norm": 0.44028225541114807, "learning_rate": 8e-05, "loss": 1.5843, "step": 7932 }, { "epoch": 0.8573435642494326, "grad_norm": 0.48781245946884155, "learning_rate": 8e-05, "loss": 1.7104, "step": 7933 }, { "epoch": 0.8574516373068194, "grad_norm": 0.4990594685077667, "learning_rate": 8e-05, "loss": 1.753, "step": 7934 }, { "epoch": 0.8575597103642062, "grad_norm": 0.4697786569595337, "learning_rate": 8e-05, "loss": 1.6785, "step": 7935 }, { "epoch": 0.857667783421593, "grad_norm": 0.4239160716533661, "learning_rate": 8e-05, "loss": 1.5731, "step": 7936 }, { "epoch": 0.8577758564789798, "grad_norm": 0.4806271195411682, "learning_rate": 8e-05, "loss": 1.7699, "step": 7937 }, { "epoch": 0.8578839295363666, "grad_norm": 0.46088525652885437, "learning_rate": 8e-05, "loss": 1.81, "step": 7938 }, { "epoch": 0.8579920025937534, "grad_norm": 0.4488910436630249, "learning_rate": 8e-05, "loss": 1.57, "step": 7939 }, { "epoch": 0.8581000756511402, "grad_norm": 0.46922391653060913, "learning_rate": 8e-05, "loss": 1.7267, "step": 7940 }, { "epoch": 0.858208148708527, "grad_norm": 0.4357219934463501, "learning_rate": 8e-05, "loss": 1.7356, "step": 7941 }, { "epoch": 0.8583162217659137, "grad_norm": 0.4788133502006531, "learning_rate": 8e-05, "loss": 1.6884, "step": 7942 }, { "epoch": 0.8584242948233005, "grad_norm": 0.4429621696472168, "learning_rate": 8e-05, "loss": 1.7735, "step": 7943 }, { "epoch": 0.8585323678806873, "grad_norm": 0.4950495958328247, "learning_rate": 8e-05, "loss": 1.7761, "step": 7944 }, { "epoch": 0.8586404409380741, "grad_norm": 0.44632941484451294, "learning_rate": 8e-05, "loss": 1.7, "step": 7945 }, { "epoch": 0.8587485139954609, "grad_norm": 0.5546655654907227, "learning_rate": 8e-05, "loss": 1.6569, "step": 7946 }, { "epoch": 0.8588565870528477, "grad_norm": 0.5066221952438354, "learning_rate": 8e-05, "loss": 1.8231, "step": 7947 }, { "epoch": 0.8589646601102345, "grad_norm": 0.4152662754058838, "learning_rate": 8e-05, "loss": 1.7878, "step": 7948 }, { "epoch": 0.8590727331676213, "grad_norm": 0.4642519950866699, "learning_rate": 8e-05, "loss": 1.7646, "step": 7949 }, { "epoch": 0.8591808062250081, "grad_norm": 0.44114601612091064, "learning_rate": 8e-05, "loss": 1.7745, "step": 7950 }, { "epoch": 0.859288879282395, "grad_norm": 0.463593065738678, "learning_rate": 8e-05, "loss": 1.8303, "step": 7951 }, { "epoch": 0.8593969523397816, "grad_norm": 0.4633841812610626, "learning_rate": 8e-05, "loss": 1.7806, "step": 7952 }, { "epoch": 0.8595050253971684, "grad_norm": 0.4635094106197357, "learning_rate": 8e-05, "loss": 1.7188, "step": 7953 }, { "epoch": 0.8596130984545552, "grad_norm": 0.44934216141700745, "learning_rate": 8e-05, "loss": 1.692, "step": 7954 }, { "epoch": 0.859721171511942, "grad_norm": 0.4419552981853485, "learning_rate": 8e-05, "loss": 1.6869, "step": 7955 }, { "epoch": 0.8598292445693289, "grad_norm": 0.45257169008255005, "learning_rate": 8e-05, "loss": 1.7052, "step": 7956 }, { "epoch": 0.8599373176267157, "grad_norm": 0.45502182841300964, "learning_rate": 8e-05, "loss": 1.8695, "step": 7957 }, { "epoch": 0.8600453906841025, "grad_norm": 0.42643094062805176, "learning_rate": 8e-05, "loss": 1.7018, "step": 7958 }, { "epoch": 0.8601534637414893, "grad_norm": 0.4450301229953766, "learning_rate": 8e-05, "loss": 1.6979, "step": 7959 }, { "epoch": 0.8602615367988761, "grad_norm": 0.4780561625957489, "learning_rate": 8e-05, "loss": 1.5949, "step": 7960 }, { "epoch": 0.8603696098562629, "grad_norm": 0.46366944909095764, "learning_rate": 8e-05, "loss": 1.6712, "step": 7961 }, { "epoch": 0.8604776829136497, "grad_norm": 0.42592066526412964, "learning_rate": 8e-05, "loss": 1.5929, "step": 7962 }, { "epoch": 0.8605857559710364, "grad_norm": 0.5179824829101562, "learning_rate": 8e-05, "loss": 1.8799, "step": 7963 }, { "epoch": 0.8606938290284232, "grad_norm": 0.42578765749931335, "learning_rate": 8e-05, "loss": 1.6376, "step": 7964 }, { "epoch": 0.86080190208581, "grad_norm": 0.5046023726463318, "learning_rate": 8e-05, "loss": 1.7387, "step": 7965 }, { "epoch": 0.8609099751431968, "grad_norm": 0.4673003554344177, "learning_rate": 8e-05, "loss": 1.6884, "step": 7966 }, { "epoch": 0.8610180482005836, "grad_norm": 0.460065633058548, "learning_rate": 8e-05, "loss": 1.7822, "step": 7967 }, { "epoch": 0.8611261212579704, "grad_norm": 0.47220608592033386, "learning_rate": 8e-05, "loss": 1.6662, "step": 7968 }, { "epoch": 0.8612341943153572, "grad_norm": 0.4081169068813324, "learning_rate": 8e-05, "loss": 1.6762, "step": 7969 }, { "epoch": 0.861342267372744, "grad_norm": 0.43150410056114197, "learning_rate": 8e-05, "loss": 1.6508, "step": 7970 }, { "epoch": 0.8614503404301308, "grad_norm": 0.44850364327430725, "learning_rate": 8e-05, "loss": 1.6486, "step": 7971 }, { "epoch": 0.8615584134875176, "grad_norm": 0.5928858518600464, "learning_rate": 8e-05, "loss": 1.9341, "step": 7972 }, { "epoch": 0.8616664865449044, "grad_norm": 0.4500442147254944, "learning_rate": 8e-05, "loss": 1.7421, "step": 7973 }, { "epoch": 0.8617745596022911, "grad_norm": 0.42417779564857483, "learning_rate": 8e-05, "loss": 1.7222, "step": 7974 }, { "epoch": 0.8618826326596779, "grad_norm": 0.4444979131221771, "learning_rate": 8e-05, "loss": 1.5436, "step": 7975 }, { "epoch": 0.8619907057170647, "grad_norm": 0.527855396270752, "learning_rate": 8e-05, "loss": 1.6854, "step": 7976 }, { "epoch": 0.8620987787744515, "grad_norm": 0.46008947491645813, "learning_rate": 8e-05, "loss": 1.6081, "step": 7977 }, { "epoch": 0.8622068518318383, "grad_norm": 0.4656948745250702, "learning_rate": 8e-05, "loss": 1.4968, "step": 7978 }, { "epoch": 0.8623149248892251, "grad_norm": 0.4979380965232849, "learning_rate": 8e-05, "loss": 1.8189, "step": 7979 }, { "epoch": 0.8624229979466119, "grad_norm": 0.5103961229324341, "learning_rate": 8e-05, "loss": 1.7228, "step": 7980 }, { "epoch": 0.8625310710039987, "grad_norm": 0.4527387320995331, "learning_rate": 8e-05, "loss": 1.8078, "step": 7981 }, { "epoch": 0.8626391440613855, "grad_norm": 0.4807199239730835, "learning_rate": 8e-05, "loss": 1.7012, "step": 7982 }, { "epoch": 0.8627472171187723, "grad_norm": 0.44702649116516113, "learning_rate": 8e-05, "loss": 1.5956, "step": 7983 }, { "epoch": 0.862855290176159, "grad_norm": 0.4788787066936493, "learning_rate": 8e-05, "loss": 1.798, "step": 7984 }, { "epoch": 0.8629633632335458, "grad_norm": 0.4571691155433655, "learning_rate": 8e-05, "loss": 1.7522, "step": 7985 }, { "epoch": 0.8630714362909326, "grad_norm": 0.44317540526390076, "learning_rate": 8e-05, "loss": 1.761, "step": 7986 }, { "epoch": 0.8631795093483194, "grad_norm": 0.43532758951187134, "learning_rate": 8e-05, "loss": 1.6379, "step": 7987 }, { "epoch": 0.8632875824057062, "grad_norm": 0.4338513910770416, "learning_rate": 8e-05, "loss": 1.6872, "step": 7988 }, { "epoch": 0.863395655463093, "grad_norm": 0.5338659882545471, "learning_rate": 8e-05, "loss": 1.9001, "step": 7989 }, { "epoch": 0.8635037285204799, "grad_norm": 0.48971208930015564, "learning_rate": 8e-05, "loss": 1.7764, "step": 7990 }, { "epoch": 0.8636118015778667, "grad_norm": 0.5139514803886414, "learning_rate": 8e-05, "loss": 1.8775, "step": 7991 }, { "epoch": 0.8637198746352535, "grad_norm": 0.4206944406032562, "learning_rate": 8e-05, "loss": 1.5938, "step": 7992 }, { "epoch": 0.8638279476926403, "grad_norm": 0.4693449139595032, "learning_rate": 8e-05, "loss": 1.8128, "step": 7993 }, { "epoch": 0.8639360207500271, "grad_norm": 0.4733228087425232, "learning_rate": 8e-05, "loss": 1.6573, "step": 7994 }, { "epoch": 0.8640440938074138, "grad_norm": 0.45364463329315186, "learning_rate": 8e-05, "loss": 1.6335, "step": 7995 }, { "epoch": 0.8641521668648006, "grad_norm": 0.4712215065956116, "learning_rate": 8e-05, "loss": 1.7145, "step": 7996 }, { "epoch": 0.8642602399221874, "grad_norm": 0.43816348910331726, "learning_rate": 8e-05, "loss": 1.5357, "step": 7997 }, { "epoch": 0.8643683129795742, "grad_norm": 0.4251561462879181, "learning_rate": 8e-05, "loss": 1.6012, "step": 7998 }, { "epoch": 0.864476386036961, "grad_norm": 0.448286235332489, "learning_rate": 8e-05, "loss": 1.652, "step": 7999 }, { "epoch": 0.8645844590943478, "grad_norm": 0.45196107029914856, "learning_rate": 8e-05, "loss": 1.6582, "step": 8000 }, { "epoch": 0.8646925321517346, "grad_norm": 0.48184409737586975, "learning_rate": 8e-05, "loss": 1.6889, "step": 8001 }, { "epoch": 0.8648006052091214, "grad_norm": 0.44312572479248047, "learning_rate": 8e-05, "loss": 1.5993, "step": 8002 }, { "epoch": 0.8649086782665082, "grad_norm": 0.4186769425868988, "learning_rate": 8e-05, "loss": 1.6831, "step": 8003 }, { "epoch": 0.865016751323895, "grad_norm": 0.4646477997303009, "learning_rate": 8e-05, "loss": 1.6901, "step": 8004 }, { "epoch": 0.8651248243812818, "grad_norm": 0.45062264800071716, "learning_rate": 8e-05, "loss": 1.6751, "step": 8005 }, { "epoch": 0.8652328974386685, "grad_norm": 0.45875242352485657, "learning_rate": 8e-05, "loss": 1.7184, "step": 8006 }, { "epoch": 0.8653409704960553, "grad_norm": 0.4841173589229584, "learning_rate": 8e-05, "loss": 1.8398, "step": 8007 }, { "epoch": 0.8654490435534421, "grad_norm": 0.42683279514312744, "learning_rate": 8e-05, "loss": 1.7133, "step": 8008 }, { "epoch": 0.8655571166108289, "grad_norm": 0.4551803469657898, "learning_rate": 8e-05, "loss": 1.5875, "step": 8009 }, { "epoch": 0.8656651896682157, "grad_norm": 0.4585631787776947, "learning_rate": 8e-05, "loss": 1.7497, "step": 8010 }, { "epoch": 0.8657732627256025, "grad_norm": 0.4488878846168518, "learning_rate": 8e-05, "loss": 1.5814, "step": 8011 }, { "epoch": 0.8658813357829893, "grad_norm": 0.4079286754131317, "learning_rate": 8e-05, "loss": 1.5697, "step": 8012 }, { "epoch": 0.8659894088403761, "grad_norm": 0.42215996980667114, "learning_rate": 8e-05, "loss": 1.5901, "step": 8013 }, { "epoch": 0.8660974818977629, "grad_norm": 0.43591803312301636, "learning_rate": 8e-05, "loss": 1.645, "step": 8014 }, { "epoch": 0.8662055549551497, "grad_norm": 0.46637967228889465, "learning_rate": 8e-05, "loss": 1.6728, "step": 8015 }, { "epoch": 0.8663136280125365, "grad_norm": 0.4329204261302948, "learning_rate": 8e-05, "loss": 1.6771, "step": 8016 }, { "epoch": 0.8664217010699232, "grad_norm": 0.4525308609008789, "learning_rate": 8e-05, "loss": 1.7087, "step": 8017 }, { "epoch": 0.86652977412731, "grad_norm": 0.48071253299713135, "learning_rate": 8e-05, "loss": 1.8392, "step": 8018 }, { "epoch": 0.8666378471846968, "grad_norm": 0.4327167570590973, "learning_rate": 8e-05, "loss": 1.6609, "step": 8019 }, { "epoch": 0.8667459202420836, "grad_norm": 0.4286437928676605, "learning_rate": 8e-05, "loss": 1.6973, "step": 8020 }, { "epoch": 0.8668539932994704, "grad_norm": 0.4899448752403259, "learning_rate": 8e-05, "loss": 1.76, "step": 8021 }, { "epoch": 0.8669620663568572, "grad_norm": 0.4950711727142334, "learning_rate": 8e-05, "loss": 1.7378, "step": 8022 }, { "epoch": 0.867070139414244, "grad_norm": 0.4294360876083374, "learning_rate": 8e-05, "loss": 1.6729, "step": 8023 }, { "epoch": 0.8671782124716308, "grad_norm": 0.41549623012542725, "learning_rate": 8e-05, "loss": 1.533, "step": 8024 }, { "epoch": 0.8672862855290177, "grad_norm": 0.4590390920639038, "learning_rate": 8e-05, "loss": 1.7546, "step": 8025 }, { "epoch": 0.8673943585864045, "grad_norm": 0.49034008383750916, "learning_rate": 8e-05, "loss": 1.9837, "step": 8026 }, { "epoch": 0.8675024316437911, "grad_norm": 0.43042251467704773, "learning_rate": 8e-05, "loss": 1.6788, "step": 8027 }, { "epoch": 0.867610504701178, "grad_norm": 0.4477149248123169, "learning_rate": 8e-05, "loss": 1.6927, "step": 8028 }, { "epoch": 0.8677185777585648, "grad_norm": 0.4566022753715515, "learning_rate": 8e-05, "loss": 1.5488, "step": 8029 }, { "epoch": 0.8678266508159516, "grad_norm": 0.4868915379047394, "learning_rate": 8e-05, "loss": 1.6721, "step": 8030 }, { "epoch": 0.8679347238733384, "grad_norm": 0.41555678844451904, "learning_rate": 8e-05, "loss": 1.7014, "step": 8031 }, { "epoch": 0.8680427969307252, "grad_norm": 0.4446346163749695, "learning_rate": 8e-05, "loss": 1.6999, "step": 8032 }, { "epoch": 0.868150869988112, "grad_norm": 0.4145359992980957, "learning_rate": 8e-05, "loss": 1.7258, "step": 8033 }, { "epoch": 0.8682589430454988, "grad_norm": 0.4600366950035095, "learning_rate": 8e-05, "loss": 1.7157, "step": 8034 }, { "epoch": 0.8683670161028856, "grad_norm": 0.4201643764972687, "learning_rate": 8e-05, "loss": 1.675, "step": 8035 }, { "epoch": 0.8684750891602724, "grad_norm": 0.4357151687145233, "learning_rate": 8e-05, "loss": 1.5969, "step": 8036 }, { "epoch": 0.8685831622176592, "grad_norm": 0.44850805401802063, "learning_rate": 8e-05, "loss": 1.6761, "step": 8037 }, { "epoch": 0.8686912352750459, "grad_norm": 0.43617212772369385, "learning_rate": 8e-05, "loss": 1.7149, "step": 8038 }, { "epoch": 0.8687993083324327, "grad_norm": 0.4720522165298462, "learning_rate": 8e-05, "loss": 1.8067, "step": 8039 }, { "epoch": 0.8689073813898195, "grad_norm": 0.4063263535499573, "learning_rate": 8e-05, "loss": 1.5362, "step": 8040 }, { "epoch": 0.8690154544472063, "grad_norm": 0.42305687069892883, "learning_rate": 8e-05, "loss": 1.7163, "step": 8041 }, { "epoch": 0.8691235275045931, "grad_norm": 0.47180938720703125, "learning_rate": 8e-05, "loss": 1.786, "step": 8042 }, { "epoch": 0.8692316005619799, "grad_norm": 0.45317959785461426, "learning_rate": 8e-05, "loss": 1.7395, "step": 8043 }, { "epoch": 0.8693396736193667, "grad_norm": 0.4483974874019623, "learning_rate": 8e-05, "loss": 1.5965, "step": 8044 }, { "epoch": 0.8694477466767535, "grad_norm": 0.45561274886131287, "learning_rate": 8e-05, "loss": 1.6252, "step": 8045 }, { "epoch": 0.8695558197341403, "grad_norm": 0.44151806831359863, "learning_rate": 8e-05, "loss": 1.7236, "step": 8046 }, { "epoch": 0.8696638927915271, "grad_norm": 0.4433801472187042, "learning_rate": 8e-05, "loss": 1.7374, "step": 8047 }, { "epoch": 0.8697719658489139, "grad_norm": 0.4297448992729187, "learning_rate": 8e-05, "loss": 1.5242, "step": 8048 }, { "epoch": 0.8698800389063006, "grad_norm": 0.4326449930667877, "learning_rate": 8e-05, "loss": 1.7153, "step": 8049 }, { "epoch": 0.8699881119636874, "grad_norm": 0.43021202087402344, "learning_rate": 8e-05, "loss": 1.7155, "step": 8050 }, { "epoch": 0.8700961850210742, "grad_norm": 0.4100959002971649, "learning_rate": 8e-05, "loss": 1.521, "step": 8051 }, { "epoch": 0.870204258078461, "grad_norm": 0.4358441233634949, "learning_rate": 8e-05, "loss": 1.7275, "step": 8052 }, { "epoch": 0.8703123311358478, "grad_norm": 0.4578995704650879, "learning_rate": 8e-05, "loss": 1.7301, "step": 8053 }, { "epoch": 0.8704204041932346, "grad_norm": 0.4341869354248047, "learning_rate": 8e-05, "loss": 1.7412, "step": 8054 }, { "epoch": 0.8705284772506214, "grad_norm": 0.44196125864982605, "learning_rate": 8e-05, "loss": 1.6223, "step": 8055 }, { "epoch": 0.8706365503080082, "grad_norm": 0.4437796473503113, "learning_rate": 8e-05, "loss": 1.7268, "step": 8056 }, { "epoch": 0.870744623365395, "grad_norm": 0.4813111424446106, "learning_rate": 8e-05, "loss": 1.863, "step": 8057 }, { "epoch": 0.8708526964227818, "grad_norm": 0.43085867166519165, "learning_rate": 8e-05, "loss": 1.5962, "step": 8058 }, { "epoch": 0.8709607694801685, "grad_norm": 0.6380683779716492, "learning_rate": 8e-05, "loss": 1.7079, "step": 8059 }, { "epoch": 0.8710688425375553, "grad_norm": 0.4531729221343994, "learning_rate": 8e-05, "loss": 1.7224, "step": 8060 }, { "epoch": 0.8711769155949421, "grad_norm": 0.4194914400577545, "learning_rate": 8e-05, "loss": 1.6178, "step": 8061 }, { "epoch": 0.871284988652329, "grad_norm": 0.4433860778808594, "learning_rate": 8e-05, "loss": 1.6926, "step": 8062 }, { "epoch": 0.8713930617097158, "grad_norm": 0.4879300594329834, "learning_rate": 8e-05, "loss": 1.7322, "step": 8063 }, { "epoch": 0.8715011347671026, "grad_norm": 0.42790040373802185, "learning_rate": 8e-05, "loss": 1.7162, "step": 8064 }, { "epoch": 0.8716092078244894, "grad_norm": 0.43462350964546204, "learning_rate": 8e-05, "loss": 1.7237, "step": 8065 }, { "epoch": 0.8717172808818762, "grad_norm": 0.4268544316291809, "learning_rate": 8e-05, "loss": 1.645, "step": 8066 }, { "epoch": 0.871825353939263, "grad_norm": 0.43259182572364807, "learning_rate": 8e-05, "loss": 1.6291, "step": 8067 }, { "epoch": 0.8719334269966498, "grad_norm": 0.462386816740036, "learning_rate": 8e-05, "loss": 1.8467, "step": 8068 }, { "epoch": 0.8720415000540366, "grad_norm": 0.4495423436164856, "learning_rate": 8e-05, "loss": 1.8593, "step": 8069 }, { "epoch": 0.8721495731114233, "grad_norm": 0.4333827495574951, "learning_rate": 8e-05, "loss": 1.7016, "step": 8070 }, { "epoch": 0.8722576461688101, "grad_norm": 0.460318922996521, "learning_rate": 8e-05, "loss": 1.6743, "step": 8071 }, { "epoch": 0.8723657192261969, "grad_norm": 0.4273485243320465, "learning_rate": 8e-05, "loss": 1.6872, "step": 8072 }, { "epoch": 0.8724737922835837, "grad_norm": 0.4488712251186371, "learning_rate": 8e-05, "loss": 1.8323, "step": 8073 }, { "epoch": 0.8725818653409705, "grad_norm": 0.4759257137775421, "learning_rate": 8e-05, "loss": 1.8406, "step": 8074 }, { "epoch": 0.8726899383983573, "grad_norm": 0.42181700468063354, "learning_rate": 8e-05, "loss": 1.6322, "step": 8075 }, { "epoch": 0.8727980114557441, "grad_norm": 0.4573400318622589, "learning_rate": 8e-05, "loss": 1.6508, "step": 8076 }, { "epoch": 0.8729060845131309, "grad_norm": 0.4416152834892273, "learning_rate": 8e-05, "loss": 1.7344, "step": 8077 }, { "epoch": 0.8730141575705177, "grad_norm": 0.3954513669013977, "learning_rate": 8e-05, "loss": 1.5134, "step": 8078 }, { "epoch": 0.8731222306279045, "grad_norm": 0.43582385778427124, "learning_rate": 8e-05, "loss": 1.6443, "step": 8079 }, { "epoch": 0.8732303036852913, "grad_norm": 0.44281578063964844, "learning_rate": 8e-05, "loss": 1.6908, "step": 8080 }, { "epoch": 0.873338376742678, "grad_norm": 0.46098071336746216, "learning_rate": 8e-05, "loss": 1.7345, "step": 8081 }, { "epoch": 0.8734464498000648, "grad_norm": 0.5036110281944275, "learning_rate": 8e-05, "loss": 1.7192, "step": 8082 }, { "epoch": 0.8735545228574516, "grad_norm": 0.4196837842464447, "learning_rate": 8e-05, "loss": 1.7442, "step": 8083 }, { "epoch": 0.8736625959148384, "grad_norm": 0.5134924054145813, "learning_rate": 8e-05, "loss": 1.7015, "step": 8084 }, { "epoch": 0.8737706689722252, "grad_norm": 0.4318130910396576, "learning_rate": 8e-05, "loss": 1.6899, "step": 8085 }, { "epoch": 0.873878742029612, "grad_norm": 0.4541942775249481, "learning_rate": 8e-05, "loss": 1.6596, "step": 8086 }, { "epoch": 0.8739868150869988, "grad_norm": 0.4097471535205841, "learning_rate": 8e-05, "loss": 1.6108, "step": 8087 }, { "epoch": 0.8740948881443856, "grad_norm": 0.4151787757873535, "learning_rate": 8e-05, "loss": 1.7098, "step": 8088 }, { "epoch": 0.8742029612017724, "grad_norm": 0.47504422068595886, "learning_rate": 8e-05, "loss": 1.8813, "step": 8089 }, { "epoch": 0.8743110342591592, "grad_norm": 0.42852380871772766, "learning_rate": 8e-05, "loss": 1.6879, "step": 8090 }, { "epoch": 0.874419107316546, "grad_norm": 0.4532714784145355, "learning_rate": 8e-05, "loss": 1.6629, "step": 8091 }, { "epoch": 0.8745271803739327, "grad_norm": 0.44209980964660645, "learning_rate": 8e-05, "loss": 1.693, "step": 8092 }, { "epoch": 0.8746352534313195, "grad_norm": 0.4422018527984619, "learning_rate": 8e-05, "loss": 1.7256, "step": 8093 }, { "epoch": 0.8747433264887063, "grad_norm": 0.45954591035842896, "learning_rate": 8e-05, "loss": 1.5981, "step": 8094 }, { "epoch": 0.8748513995460931, "grad_norm": 0.4552217125892639, "learning_rate": 8e-05, "loss": 1.5037, "step": 8095 }, { "epoch": 0.87495947260348, "grad_norm": 0.4683581590652466, "learning_rate": 8e-05, "loss": 1.6961, "step": 8096 }, { "epoch": 0.8750675456608668, "grad_norm": 0.4273242950439453, "learning_rate": 8e-05, "loss": 1.5559, "step": 8097 }, { "epoch": 0.8751756187182536, "grad_norm": 0.48293766379356384, "learning_rate": 8e-05, "loss": 1.9065, "step": 8098 }, { "epoch": 0.8752836917756404, "grad_norm": 0.4634474515914917, "learning_rate": 8e-05, "loss": 1.6566, "step": 8099 }, { "epoch": 0.8753917648330272, "grad_norm": 0.45031338930130005, "learning_rate": 8e-05, "loss": 1.6478, "step": 8100 }, { "epoch": 0.875499837890414, "grad_norm": 0.4681468605995178, "learning_rate": 8e-05, "loss": 1.5926, "step": 8101 }, { "epoch": 0.8756079109478007, "grad_norm": 0.429359495639801, "learning_rate": 8e-05, "loss": 1.7382, "step": 8102 }, { "epoch": 0.8757159840051875, "grad_norm": 0.4216657280921936, "learning_rate": 8e-05, "loss": 1.6251, "step": 8103 }, { "epoch": 0.8758240570625743, "grad_norm": 0.462704062461853, "learning_rate": 8e-05, "loss": 1.6094, "step": 8104 }, { "epoch": 0.8759321301199611, "grad_norm": 0.47024309635162354, "learning_rate": 8e-05, "loss": 1.7538, "step": 8105 }, { "epoch": 0.8760402031773479, "grad_norm": 0.45079654455184937, "learning_rate": 8e-05, "loss": 1.6844, "step": 8106 }, { "epoch": 0.8761482762347347, "grad_norm": 0.4356571137905121, "learning_rate": 8e-05, "loss": 1.6942, "step": 8107 }, { "epoch": 0.8762563492921215, "grad_norm": 0.43005576729774475, "learning_rate": 8e-05, "loss": 1.7182, "step": 8108 }, { "epoch": 0.8763644223495083, "grad_norm": 0.42737293243408203, "learning_rate": 8e-05, "loss": 1.6909, "step": 8109 }, { "epoch": 0.8764724954068951, "grad_norm": 0.4608423113822937, "learning_rate": 8e-05, "loss": 1.6491, "step": 8110 }, { "epoch": 0.8765805684642819, "grad_norm": 0.5392202734947205, "learning_rate": 8e-05, "loss": 1.7516, "step": 8111 }, { "epoch": 0.8766886415216687, "grad_norm": 0.4717053771018982, "learning_rate": 8e-05, "loss": 1.796, "step": 8112 }, { "epoch": 0.8767967145790554, "grad_norm": 0.4398230314254761, "learning_rate": 8e-05, "loss": 1.5665, "step": 8113 }, { "epoch": 0.8769047876364422, "grad_norm": 0.47309714555740356, "learning_rate": 8e-05, "loss": 1.7958, "step": 8114 }, { "epoch": 0.877012860693829, "grad_norm": 0.42239314317703247, "learning_rate": 8e-05, "loss": 1.6207, "step": 8115 }, { "epoch": 0.8771209337512158, "grad_norm": 0.42778778076171875, "learning_rate": 8e-05, "loss": 1.7163, "step": 8116 }, { "epoch": 0.8772290068086026, "grad_norm": 0.4372841715812683, "learning_rate": 8e-05, "loss": 1.6874, "step": 8117 }, { "epoch": 0.8773370798659894, "grad_norm": 0.46416017413139343, "learning_rate": 8e-05, "loss": 1.6394, "step": 8118 }, { "epoch": 0.8774451529233762, "grad_norm": 0.44732409715652466, "learning_rate": 8e-05, "loss": 1.6955, "step": 8119 }, { "epoch": 0.877553225980763, "grad_norm": 0.4428260326385498, "learning_rate": 8e-05, "loss": 1.5694, "step": 8120 }, { "epoch": 0.8776612990381498, "grad_norm": 0.4169573187828064, "learning_rate": 8e-05, "loss": 1.7854, "step": 8121 }, { "epoch": 0.8777693720955366, "grad_norm": 0.4257272183895111, "learning_rate": 8e-05, "loss": 1.6864, "step": 8122 }, { "epoch": 0.8778774451529234, "grad_norm": 0.4339004456996918, "learning_rate": 8e-05, "loss": 1.5897, "step": 8123 }, { "epoch": 0.8779855182103101, "grad_norm": 0.44348061084747314, "learning_rate": 8e-05, "loss": 1.6927, "step": 8124 }, { "epoch": 0.8780935912676969, "grad_norm": 0.4787020981311798, "learning_rate": 8e-05, "loss": 1.7456, "step": 8125 }, { "epoch": 0.8782016643250837, "grad_norm": 0.4301522374153137, "learning_rate": 8e-05, "loss": 1.4817, "step": 8126 }, { "epoch": 0.8783097373824705, "grad_norm": 0.4229406714439392, "learning_rate": 8e-05, "loss": 1.6672, "step": 8127 }, { "epoch": 0.8784178104398573, "grad_norm": 0.4403367340564728, "learning_rate": 8e-05, "loss": 1.6397, "step": 8128 }, { "epoch": 0.8785258834972441, "grad_norm": 0.4349992573261261, "learning_rate": 8e-05, "loss": 1.7504, "step": 8129 }, { "epoch": 0.878633956554631, "grad_norm": 0.44676923751831055, "learning_rate": 8e-05, "loss": 1.9034, "step": 8130 }, { "epoch": 0.8787420296120177, "grad_norm": 0.5329145193099976, "learning_rate": 8e-05, "loss": 1.8201, "step": 8131 }, { "epoch": 0.8788501026694046, "grad_norm": 0.4742286503314972, "learning_rate": 8e-05, "loss": 1.812, "step": 8132 }, { "epoch": 0.8789581757267914, "grad_norm": 0.47718551754951477, "learning_rate": 8e-05, "loss": 1.7446, "step": 8133 }, { "epoch": 0.879066248784178, "grad_norm": 0.42435798048973083, "learning_rate": 8e-05, "loss": 1.7018, "step": 8134 }, { "epoch": 0.8791743218415649, "grad_norm": 0.47154441475868225, "learning_rate": 8e-05, "loss": 1.7201, "step": 8135 }, { "epoch": 0.8792823948989517, "grad_norm": 0.5019599795341492, "learning_rate": 8e-05, "loss": 1.6856, "step": 8136 }, { "epoch": 0.8793904679563385, "grad_norm": 0.44317007064819336, "learning_rate": 8e-05, "loss": 1.7839, "step": 8137 }, { "epoch": 0.8794985410137253, "grad_norm": 0.42690062522888184, "learning_rate": 8e-05, "loss": 1.7559, "step": 8138 }, { "epoch": 0.8796066140711121, "grad_norm": 0.4147360622882843, "learning_rate": 8e-05, "loss": 1.5675, "step": 8139 }, { "epoch": 0.8797146871284989, "grad_norm": 0.4118559956550598, "learning_rate": 8e-05, "loss": 1.7034, "step": 8140 }, { "epoch": 0.8798227601858857, "grad_norm": 0.41136324405670166, "learning_rate": 8e-05, "loss": 1.4932, "step": 8141 }, { "epoch": 0.8799308332432725, "grad_norm": 0.4697949290275574, "learning_rate": 8e-05, "loss": 1.6451, "step": 8142 }, { "epoch": 0.8800389063006593, "grad_norm": 0.45062047243118286, "learning_rate": 8e-05, "loss": 1.6409, "step": 8143 }, { "epoch": 0.8801469793580461, "grad_norm": 0.4443132281303406, "learning_rate": 8e-05, "loss": 1.8325, "step": 8144 }, { "epoch": 0.8802550524154328, "grad_norm": 0.46742191910743713, "learning_rate": 8e-05, "loss": 1.7496, "step": 8145 }, { "epoch": 0.8803631254728196, "grad_norm": 0.43261972069740295, "learning_rate": 8e-05, "loss": 1.5522, "step": 8146 }, { "epoch": 0.8804711985302064, "grad_norm": 0.43583792448043823, "learning_rate": 8e-05, "loss": 1.5417, "step": 8147 }, { "epoch": 0.8805792715875932, "grad_norm": 0.47855275869369507, "learning_rate": 8e-05, "loss": 1.7189, "step": 8148 }, { "epoch": 0.88068734464498, "grad_norm": 0.4082844853401184, "learning_rate": 8e-05, "loss": 1.634, "step": 8149 }, { "epoch": 0.8807954177023668, "grad_norm": 0.4381188750267029, "learning_rate": 8e-05, "loss": 1.6544, "step": 8150 }, { "epoch": 0.8809034907597536, "grad_norm": 0.46315550804138184, "learning_rate": 8e-05, "loss": 1.7513, "step": 8151 }, { "epoch": 0.8810115638171404, "grad_norm": 0.43444299697875977, "learning_rate": 8e-05, "loss": 1.6806, "step": 8152 }, { "epoch": 0.8811196368745272, "grad_norm": 0.45370545983314514, "learning_rate": 8e-05, "loss": 1.6712, "step": 8153 }, { "epoch": 0.881227709931914, "grad_norm": 0.447513222694397, "learning_rate": 8e-05, "loss": 1.7007, "step": 8154 }, { "epoch": 0.8813357829893008, "grad_norm": 0.4934902489185333, "learning_rate": 8e-05, "loss": 1.833, "step": 8155 }, { "epoch": 0.8814438560466875, "grad_norm": 0.4226034879684448, "learning_rate": 8e-05, "loss": 1.5114, "step": 8156 }, { "epoch": 0.8815519291040743, "grad_norm": 0.44159528613090515, "learning_rate": 8e-05, "loss": 1.7176, "step": 8157 }, { "epoch": 0.8816600021614611, "grad_norm": 0.45316073298454285, "learning_rate": 8e-05, "loss": 1.8265, "step": 8158 }, { "epoch": 0.8817680752188479, "grad_norm": 0.4743443429470062, "learning_rate": 8e-05, "loss": 1.6963, "step": 8159 }, { "epoch": 0.8818761482762347, "grad_norm": 0.48674991726875305, "learning_rate": 8e-05, "loss": 1.9294, "step": 8160 }, { "epoch": 0.8819842213336215, "grad_norm": 0.5015445351600647, "learning_rate": 8e-05, "loss": 1.8467, "step": 8161 }, { "epoch": 0.8820922943910083, "grad_norm": 0.4326673746109009, "learning_rate": 8e-05, "loss": 1.6865, "step": 8162 }, { "epoch": 0.8822003674483951, "grad_norm": 0.47976547479629517, "learning_rate": 8e-05, "loss": 1.841, "step": 8163 }, { "epoch": 0.8823084405057819, "grad_norm": 0.4426446855068207, "learning_rate": 8e-05, "loss": 1.7062, "step": 8164 }, { "epoch": 0.8824165135631687, "grad_norm": 0.4276278614997864, "learning_rate": 8e-05, "loss": 1.6175, "step": 8165 }, { "epoch": 0.8825245866205554, "grad_norm": 0.4527157247066498, "learning_rate": 8e-05, "loss": 1.7804, "step": 8166 }, { "epoch": 0.8826326596779422, "grad_norm": 0.41782090067863464, "learning_rate": 8e-05, "loss": 1.4979, "step": 8167 }, { "epoch": 0.882740732735329, "grad_norm": 0.42873358726501465, "learning_rate": 8e-05, "loss": 1.6807, "step": 8168 }, { "epoch": 0.8828488057927159, "grad_norm": 0.435696005821228, "learning_rate": 8e-05, "loss": 1.5551, "step": 8169 }, { "epoch": 0.8829568788501027, "grad_norm": 0.4814722537994385, "learning_rate": 8e-05, "loss": 1.7145, "step": 8170 }, { "epoch": 0.8830649519074895, "grad_norm": 0.44113612174987793, "learning_rate": 8e-05, "loss": 1.6452, "step": 8171 }, { "epoch": 0.8831730249648763, "grad_norm": 0.42880779504776, "learning_rate": 8e-05, "loss": 1.718, "step": 8172 }, { "epoch": 0.8832810980222631, "grad_norm": 0.4524756371974945, "learning_rate": 8e-05, "loss": 1.598, "step": 8173 }, { "epoch": 0.8833891710796499, "grad_norm": 0.49106544256210327, "learning_rate": 8e-05, "loss": 1.8958, "step": 8174 }, { "epoch": 0.8834972441370367, "grad_norm": 0.47517967224121094, "learning_rate": 8e-05, "loss": 1.8156, "step": 8175 }, { "epoch": 0.8836053171944235, "grad_norm": 0.422445148229599, "learning_rate": 8e-05, "loss": 1.5156, "step": 8176 }, { "epoch": 0.8837133902518102, "grad_norm": 0.44621819257736206, "learning_rate": 8e-05, "loss": 1.6608, "step": 8177 }, { "epoch": 0.883821463309197, "grad_norm": 0.44670793414115906, "learning_rate": 8e-05, "loss": 1.5741, "step": 8178 }, { "epoch": 0.8839295363665838, "grad_norm": 0.45495909452438354, "learning_rate": 8e-05, "loss": 1.7236, "step": 8179 }, { "epoch": 0.8840376094239706, "grad_norm": 0.42010968923568726, "learning_rate": 8e-05, "loss": 1.6099, "step": 8180 }, { "epoch": 0.8841456824813574, "grad_norm": 0.4310305118560791, "learning_rate": 8e-05, "loss": 1.6958, "step": 8181 }, { "epoch": 0.8842537555387442, "grad_norm": 0.4557233452796936, "learning_rate": 8e-05, "loss": 1.8776, "step": 8182 }, { "epoch": 0.884361828596131, "grad_norm": 0.43670180439949036, "learning_rate": 8e-05, "loss": 1.6121, "step": 8183 }, { "epoch": 0.8844699016535178, "grad_norm": 0.4884210526943207, "learning_rate": 8e-05, "loss": 1.7138, "step": 8184 }, { "epoch": 0.8845779747109046, "grad_norm": 0.4230082333087921, "learning_rate": 8e-05, "loss": 1.5836, "step": 8185 }, { "epoch": 0.8846860477682914, "grad_norm": 0.44733700156211853, "learning_rate": 8e-05, "loss": 1.801, "step": 8186 }, { "epoch": 0.8847941208256782, "grad_norm": 0.4211227297782898, "learning_rate": 8e-05, "loss": 1.5549, "step": 8187 }, { "epoch": 0.8849021938830649, "grad_norm": 0.4236529767513275, "learning_rate": 8e-05, "loss": 1.6887, "step": 8188 }, { "epoch": 0.8850102669404517, "grad_norm": 0.4726375937461853, "learning_rate": 8e-05, "loss": 1.6949, "step": 8189 }, { "epoch": 0.8851183399978385, "grad_norm": 0.43291041254997253, "learning_rate": 8e-05, "loss": 1.7162, "step": 8190 }, { "epoch": 0.8852264130552253, "grad_norm": 0.45013728737831116, "learning_rate": 8e-05, "loss": 1.6228, "step": 8191 }, { "epoch": 0.8853344861126121, "grad_norm": 0.4211331605911255, "learning_rate": 8e-05, "loss": 1.737, "step": 8192 }, { "epoch": 0.8854425591699989, "grad_norm": 0.43857720494270325, "learning_rate": 8e-05, "loss": 1.6437, "step": 8193 }, { "epoch": 0.8855506322273857, "grad_norm": 0.42208215594291687, "learning_rate": 8e-05, "loss": 1.6197, "step": 8194 }, { "epoch": 0.8856587052847725, "grad_norm": 0.4605594575405121, "learning_rate": 8e-05, "loss": 1.7928, "step": 8195 }, { "epoch": 0.8857667783421593, "grad_norm": 0.4562883675098419, "learning_rate": 8e-05, "loss": 1.7189, "step": 8196 }, { "epoch": 0.8858748513995461, "grad_norm": 0.4960337281227112, "learning_rate": 8e-05, "loss": 1.8006, "step": 8197 }, { "epoch": 0.8859829244569329, "grad_norm": 0.4527091085910797, "learning_rate": 8e-05, "loss": 1.7675, "step": 8198 }, { "epoch": 0.8860909975143196, "grad_norm": 0.4375815689563751, "learning_rate": 8e-05, "loss": 1.6763, "step": 8199 }, { "epoch": 0.8861990705717064, "grad_norm": 0.440038800239563, "learning_rate": 8e-05, "loss": 1.7274, "step": 8200 }, { "epoch": 0.8863071436290932, "grad_norm": 0.44584304094314575, "learning_rate": 8e-05, "loss": 1.7581, "step": 8201 }, { "epoch": 0.88641521668648, "grad_norm": 0.4358159303665161, "learning_rate": 8e-05, "loss": 1.5826, "step": 8202 }, { "epoch": 0.8865232897438668, "grad_norm": 0.4416894316673279, "learning_rate": 8e-05, "loss": 1.7604, "step": 8203 }, { "epoch": 0.8866313628012537, "grad_norm": 0.43985238671302795, "learning_rate": 8e-05, "loss": 1.7087, "step": 8204 }, { "epoch": 0.8867394358586405, "grad_norm": 0.46702978014945984, "learning_rate": 8e-05, "loss": 1.6635, "step": 8205 }, { "epoch": 0.8868475089160273, "grad_norm": 0.47947898507118225, "learning_rate": 8e-05, "loss": 1.6293, "step": 8206 }, { "epoch": 0.8869555819734141, "grad_norm": 0.4409225881099701, "learning_rate": 8e-05, "loss": 1.7526, "step": 8207 }, { "epoch": 0.8870636550308009, "grad_norm": 0.42457592487335205, "learning_rate": 8e-05, "loss": 1.6468, "step": 8208 }, { "epoch": 0.8871717280881876, "grad_norm": 0.4179283678531647, "learning_rate": 8e-05, "loss": 1.611, "step": 8209 }, { "epoch": 0.8872798011455744, "grad_norm": 0.43699175119400024, "learning_rate": 8e-05, "loss": 1.6823, "step": 8210 }, { "epoch": 0.8873878742029612, "grad_norm": 0.4710932672023773, "learning_rate": 8e-05, "loss": 1.6813, "step": 8211 }, { "epoch": 0.887495947260348, "grad_norm": 0.4362621605396271, "learning_rate": 8e-05, "loss": 1.5401, "step": 8212 }, { "epoch": 0.8876040203177348, "grad_norm": 0.45291227102279663, "learning_rate": 8e-05, "loss": 1.6658, "step": 8213 }, { "epoch": 0.8877120933751216, "grad_norm": 0.4449228048324585, "learning_rate": 8e-05, "loss": 1.6593, "step": 8214 }, { "epoch": 0.8878201664325084, "grad_norm": 0.45055434107780457, "learning_rate": 8e-05, "loss": 1.5066, "step": 8215 }, { "epoch": 0.8879282394898952, "grad_norm": 0.49830082058906555, "learning_rate": 8e-05, "loss": 1.5675, "step": 8216 }, { "epoch": 0.888036312547282, "grad_norm": 0.48951008915901184, "learning_rate": 8e-05, "loss": 1.7961, "step": 8217 }, { "epoch": 0.8881443856046688, "grad_norm": 0.44039827585220337, "learning_rate": 8e-05, "loss": 1.7232, "step": 8218 }, { "epoch": 0.8882524586620556, "grad_norm": 0.41605624556541443, "learning_rate": 8e-05, "loss": 1.5519, "step": 8219 }, { "epoch": 0.8883605317194423, "grad_norm": 0.4307696223258972, "learning_rate": 8e-05, "loss": 1.7477, "step": 8220 }, { "epoch": 0.8884686047768291, "grad_norm": 0.45889967679977417, "learning_rate": 8e-05, "loss": 1.7249, "step": 8221 }, { "epoch": 0.8885766778342159, "grad_norm": 0.4311768114566803, "learning_rate": 8e-05, "loss": 1.6652, "step": 8222 }, { "epoch": 0.8886847508916027, "grad_norm": 0.43806108832359314, "learning_rate": 8e-05, "loss": 1.8169, "step": 8223 }, { "epoch": 0.8887928239489895, "grad_norm": 0.4011405408382416, "learning_rate": 8e-05, "loss": 1.5216, "step": 8224 }, { "epoch": 0.8889008970063763, "grad_norm": 0.4289395213127136, "learning_rate": 8e-05, "loss": 1.7006, "step": 8225 }, { "epoch": 0.8890089700637631, "grad_norm": 0.4899025857448578, "learning_rate": 8e-05, "loss": 1.7234, "step": 8226 }, { "epoch": 0.8891170431211499, "grad_norm": 0.4028613865375519, "learning_rate": 8e-05, "loss": 1.5506, "step": 8227 }, { "epoch": 0.8892251161785367, "grad_norm": 0.43536123633384705, "learning_rate": 8e-05, "loss": 1.5984, "step": 8228 }, { "epoch": 0.8893331892359235, "grad_norm": 0.4141228199005127, "learning_rate": 8e-05, "loss": 1.5998, "step": 8229 }, { "epoch": 0.8894412622933103, "grad_norm": 0.41170215606689453, "learning_rate": 8e-05, "loss": 1.6658, "step": 8230 }, { "epoch": 0.889549335350697, "grad_norm": 0.4261741042137146, "learning_rate": 8e-05, "loss": 1.6122, "step": 8231 }, { "epoch": 0.8896574084080838, "grad_norm": 0.4764125943183899, "learning_rate": 8e-05, "loss": 1.7749, "step": 8232 }, { "epoch": 0.8897654814654706, "grad_norm": 0.47351759672164917, "learning_rate": 8e-05, "loss": 1.7925, "step": 8233 }, { "epoch": 0.8898735545228574, "grad_norm": 0.41952645778656006, "learning_rate": 8e-05, "loss": 1.6156, "step": 8234 }, { "epoch": 0.8899816275802442, "grad_norm": 0.45828503370285034, "learning_rate": 8e-05, "loss": 1.7549, "step": 8235 }, { "epoch": 0.890089700637631, "grad_norm": 0.41890057921409607, "learning_rate": 8e-05, "loss": 1.6201, "step": 8236 }, { "epoch": 0.8901977736950178, "grad_norm": 0.43919646739959717, "learning_rate": 8e-05, "loss": 1.6782, "step": 8237 }, { "epoch": 0.8903058467524047, "grad_norm": 0.43162983655929565, "learning_rate": 8e-05, "loss": 1.5771, "step": 8238 }, { "epoch": 0.8904139198097915, "grad_norm": 0.4849119782447815, "learning_rate": 8e-05, "loss": 1.8048, "step": 8239 }, { "epoch": 0.8905219928671783, "grad_norm": 0.43947911262512207, "learning_rate": 8e-05, "loss": 1.7265, "step": 8240 }, { "epoch": 0.890630065924565, "grad_norm": 0.4542270302772522, "learning_rate": 8e-05, "loss": 1.742, "step": 8241 }, { "epoch": 0.8907381389819518, "grad_norm": 0.4155289828777313, "learning_rate": 8e-05, "loss": 1.6997, "step": 8242 }, { "epoch": 0.8908462120393386, "grad_norm": 0.45344096422195435, "learning_rate": 8e-05, "loss": 1.6356, "step": 8243 }, { "epoch": 0.8909542850967254, "grad_norm": 0.46899497509002686, "learning_rate": 8e-05, "loss": 1.7898, "step": 8244 }, { "epoch": 0.8910623581541122, "grad_norm": 0.436616450548172, "learning_rate": 8e-05, "loss": 1.4815, "step": 8245 }, { "epoch": 0.891170431211499, "grad_norm": 0.4886411726474762, "learning_rate": 8e-05, "loss": 1.8557, "step": 8246 }, { "epoch": 0.8912785042688858, "grad_norm": 0.4961124658584595, "learning_rate": 8e-05, "loss": 1.8456, "step": 8247 }, { "epoch": 0.8913865773262726, "grad_norm": 0.46739810705184937, "learning_rate": 8e-05, "loss": 1.7317, "step": 8248 }, { "epoch": 0.8914946503836594, "grad_norm": 0.49255895614624023, "learning_rate": 8e-05, "loss": 1.7406, "step": 8249 }, { "epoch": 0.8916027234410462, "grad_norm": 0.5198391079902649, "learning_rate": 8e-05, "loss": 1.8406, "step": 8250 }, { "epoch": 0.891710796498433, "grad_norm": 0.42283087968826294, "learning_rate": 8e-05, "loss": 1.4843, "step": 8251 }, { "epoch": 0.8918188695558197, "grad_norm": 0.42337098717689514, "learning_rate": 8e-05, "loss": 1.5755, "step": 8252 }, { "epoch": 0.8919269426132065, "grad_norm": 0.44624900817871094, "learning_rate": 8e-05, "loss": 1.6804, "step": 8253 }, { "epoch": 0.8920350156705933, "grad_norm": 0.42806312441825867, "learning_rate": 8e-05, "loss": 1.7226, "step": 8254 }, { "epoch": 0.8921430887279801, "grad_norm": 0.4483315944671631, "learning_rate": 8e-05, "loss": 1.6209, "step": 8255 }, { "epoch": 0.8922511617853669, "grad_norm": 0.4360932409763336, "learning_rate": 8e-05, "loss": 1.5637, "step": 8256 }, { "epoch": 0.8923592348427537, "grad_norm": 0.4841107130050659, "learning_rate": 8e-05, "loss": 1.941, "step": 8257 }, { "epoch": 0.8924673079001405, "grad_norm": 0.4441544711589813, "learning_rate": 8e-05, "loss": 1.7101, "step": 8258 }, { "epoch": 0.8925753809575273, "grad_norm": 0.4380033314228058, "learning_rate": 8e-05, "loss": 1.7465, "step": 8259 }, { "epoch": 0.8926834540149141, "grad_norm": 0.45760586857795715, "learning_rate": 8e-05, "loss": 1.8588, "step": 8260 }, { "epoch": 0.8927915270723009, "grad_norm": 0.4720771312713623, "learning_rate": 8e-05, "loss": 1.7967, "step": 8261 }, { "epoch": 0.8928996001296877, "grad_norm": 0.4738767743110657, "learning_rate": 8e-05, "loss": 1.719, "step": 8262 }, { "epoch": 0.8930076731870744, "grad_norm": 0.46117880940437317, "learning_rate": 8e-05, "loss": 1.7185, "step": 8263 }, { "epoch": 0.8931157462444612, "grad_norm": 0.4077828824520111, "learning_rate": 8e-05, "loss": 1.4991, "step": 8264 }, { "epoch": 0.893223819301848, "grad_norm": 0.4470388889312744, "learning_rate": 8e-05, "loss": 1.7436, "step": 8265 }, { "epoch": 0.8933318923592348, "grad_norm": 0.4465998113155365, "learning_rate": 8e-05, "loss": 1.6797, "step": 8266 }, { "epoch": 0.8934399654166216, "grad_norm": 0.4062787592411041, "learning_rate": 8e-05, "loss": 1.5798, "step": 8267 }, { "epoch": 0.8935480384740084, "grad_norm": 0.41225194931030273, "learning_rate": 8e-05, "loss": 1.6319, "step": 8268 }, { "epoch": 0.8936561115313952, "grad_norm": 0.44410422444343567, "learning_rate": 8e-05, "loss": 1.6769, "step": 8269 }, { "epoch": 0.893764184588782, "grad_norm": 0.41476041078567505, "learning_rate": 8e-05, "loss": 1.6161, "step": 8270 }, { "epoch": 0.8938722576461688, "grad_norm": 0.45695579051971436, "learning_rate": 8e-05, "loss": 1.362, "step": 8271 }, { "epoch": 0.8939803307035556, "grad_norm": 0.42568162083625793, "learning_rate": 8e-05, "loss": 1.5983, "step": 8272 }, { "epoch": 0.8940884037609425, "grad_norm": 0.4371042251586914, "learning_rate": 8e-05, "loss": 1.4529, "step": 8273 }, { "epoch": 0.8941964768183291, "grad_norm": 0.43588685989379883, "learning_rate": 8e-05, "loss": 1.7323, "step": 8274 }, { "epoch": 0.894304549875716, "grad_norm": 0.49097710847854614, "learning_rate": 8e-05, "loss": 1.624, "step": 8275 }, { "epoch": 0.8944126229331028, "grad_norm": 0.42733946442604065, "learning_rate": 8e-05, "loss": 1.6163, "step": 8276 }, { "epoch": 0.8945206959904896, "grad_norm": 0.4293173551559448, "learning_rate": 8e-05, "loss": 1.6795, "step": 8277 }, { "epoch": 0.8946287690478764, "grad_norm": 0.4329582452774048, "learning_rate": 8e-05, "loss": 1.5851, "step": 8278 }, { "epoch": 0.8947368421052632, "grad_norm": 0.47396793961524963, "learning_rate": 8e-05, "loss": 1.8291, "step": 8279 }, { "epoch": 0.89484491516265, "grad_norm": 0.504876434803009, "learning_rate": 8e-05, "loss": 1.7826, "step": 8280 }, { "epoch": 0.8949529882200368, "grad_norm": 0.5196802020072937, "learning_rate": 8e-05, "loss": 1.6128, "step": 8281 }, { "epoch": 0.8950610612774236, "grad_norm": 0.4381680190563202, "learning_rate": 8e-05, "loss": 1.5857, "step": 8282 }, { "epoch": 0.8951691343348104, "grad_norm": 0.44163739681243896, "learning_rate": 8e-05, "loss": 1.6497, "step": 8283 }, { "epoch": 0.8952772073921971, "grad_norm": 0.47409504652023315, "learning_rate": 8e-05, "loss": 1.837, "step": 8284 }, { "epoch": 0.8953852804495839, "grad_norm": 0.44484061002731323, "learning_rate": 8e-05, "loss": 1.6613, "step": 8285 }, { "epoch": 0.8954933535069707, "grad_norm": 0.4700358510017395, "learning_rate": 8e-05, "loss": 1.5736, "step": 8286 }, { "epoch": 0.8956014265643575, "grad_norm": 0.423698753118515, "learning_rate": 8e-05, "loss": 1.5964, "step": 8287 }, { "epoch": 0.8957094996217443, "grad_norm": 0.4546797275543213, "learning_rate": 8e-05, "loss": 1.7216, "step": 8288 }, { "epoch": 0.8958175726791311, "grad_norm": 0.4212090075016022, "learning_rate": 8e-05, "loss": 1.5911, "step": 8289 }, { "epoch": 0.8959256457365179, "grad_norm": 0.438262403011322, "learning_rate": 8e-05, "loss": 1.7139, "step": 8290 }, { "epoch": 0.8960337187939047, "grad_norm": 0.4906594753265381, "learning_rate": 8e-05, "loss": 1.6492, "step": 8291 }, { "epoch": 0.8961417918512915, "grad_norm": 0.46878743171691895, "learning_rate": 8e-05, "loss": 1.5254, "step": 8292 }, { "epoch": 0.8962498649086783, "grad_norm": 0.4880736172199249, "learning_rate": 8e-05, "loss": 1.5256, "step": 8293 }, { "epoch": 0.8963579379660651, "grad_norm": 0.4989790916442871, "learning_rate": 8e-05, "loss": 1.8546, "step": 8294 }, { "epoch": 0.8964660110234518, "grad_norm": 0.4705473482608795, "learning_rate": 8e-05, "loss": 1.7248, "step": 8295 }, { "epoch": 0.8965740840808386, "grad_norm": 0.43349066376686096, "learning_rate": 8e-05, "loss": 1.7477, "step": 8296 }, { "epoch": 0.8966821571382254, "grad_norm": 0.43323060870170593, "learning_rate": 8e-05, "loss": 1.6717, "step": 8297 }, { "epoch": 0.8967902301956122, "grad_norm": 0.490130752325058, "learning_rate": 8e-05, "loss": 1.8096, "step": 8298 }, { "epoch": 0.896898303252999, "grad_norm": 0.46816903352737427, "learning_rate": 8e-05, "loss": 1.7968, "step": 8299 }, { "epoch": 0.8970063763103858, "grad_norm": 0.4386681020259857, "learning_rate": 8e-05, "loss": 1.6591, "step": 8300 }, { "epoch": 0.8971144493677726, "grad_norm": 0.4726282060146332, "learning_rate": 8e-05, "loss": 1.6078, "step": 8301 }, { "epoch": 0.8972225224251594, "grad_norm": 0.44297823309898376, "learning_rate": 8e-05, "loss": 1.5969, "step": 8302 }, { "epoch": 0.8973305954825462, "grad_norm": 0.4665282964706421, "learning_rate": 8e-05, "loss": 1.4942, "step": 8303 }, { "epoch": 0.897438668539933, "grad_norm": 0.42639216780662537, "learning_rate": 8e-05, "loss": 1.5277, "step": 8304 }, { "epoch": 0.8975467415973198, "grad_norm": 0.4322594702243805, "learning_rate": 8e-05, "loss": 1.6265, "step": 8305 }, { "epoch": 0.8976548146547065, "grad_norm": 0.4795089364051819, "learning_rate": 8e-05, "loss": 1.6968, "step": 8306 }, { "epoch": 0.8977628877120933, "grad_norm": 0.40509361028671265, "learning_rate": 8e-05, "loss": 1.6183, "step": 8307 }, { "epoch": 0.8978709607694801, "grad_norm": 0.4683281183242798, "learning_rate": 8e-05, "loss": 1.7201, "step": 8308 }, { "epoch": 0.897979033826867, "grad_norm": 0.4598444104194641, "learning_rate": 8e-05, "loss": 1.8119, "step": 8309 }, { "epoch": 0.8980871068842537, "grad_norm": 0.42771726846694946, "learning_rate": 8e-05, "loss": 1.7149, "step": 8310 }, { "epoch": 0.8981951799416406, "grad_norm": 0.46234065294265747, "learning_rate": 8e-05, "loss": 1.6174, "step": 8311 }, { "epoch": 0.8983032529990274, "grad_norm": 0.4202989339828491, "learning_rate": 8e-05, "loss": 1.5356, "step": 8312 }, { "epoch": 0.8984113260564142, "grad_norm": 0.42098701000213623, "learning_rate": 8e-05, "loss": 1.4748, "step": 8313 }, { "epoch": 0.898519399113801, "grad_norm": 0.4147082269191742, "learning_rate": 8e-05, "loss": 1.576, "step": 8314 }, { "epoch": 0.8986274721711878, "grad_norm": 0.48106107115745544, "learning_rate": 8e-05, "loss": 1.7447, "step": 8315 }, { "epoch": 0.8987355452285745, "grad_norm": 0.4340192675590515, "learning_rate": 8e-05, "loss": 1.5703, "step": 8316 }, { "epoch": 0.8988436182859613, "grad_norm": 0.4916112720966339, "learning_rate": 8e-05, "loss": 1.8677, "step": 8317 }, { "epoch": 0.8989516913433481, "grad_norm": 0.4384276866912842, "learning_rate": 8e-05, "loss": 1.595, "step": 8318 }, { "epoch": 0.8990597644007349, "grad_norm": 0.47400811314582825, "learning_rate": 8e-05, "loss": 1.764, "step": 8319 }, { "epoch": 0.8991678374581217, "grad_norm": 0.45001474022865295, "learning_rate": 8e-05, "loss": 1.7143, "step": 8320 }, { "epoch": 0.8992759105155085, "grad_norm": 0.4930223524570465, "learning_rate": 8e-05, "loss": 1.9232, "step": 8321 }, { "epoch": 0.8993839835728953, "grad_norm": 0.44294852018356323, "learning_rate": 8e-05, "loss": 1.5625, "step": 8322 }, { "epoch": 0.8994920566302821, "grad_norm": 0.48954272270202637, "learning_rate": 8e-05, "loss": 1.6799, "step": 8323 }, { "epoch": 0.8996001296876689, "grad_norm": 0.47966060042381287, "learning_rate": 8e-05, "loss": 1.7165, "step": 8324 }, { "epoch": 0.8997082027450557, "grad_norm": 0.48302122950553894, "learning_rate": 8e-05, "loss": 1.7734, "step": 8325 }, { "epoch": 0.8998162758024425, "grad_norm": 0.45214223861694336, "learning_rate": 8e-05, "loss": 1.6182, "step": 8326 }, { "epoch": 0.8999243488598292, "grad_norm": 0.45901548862457275, "learning_rate": 8e-05, "loss": 1.7868, "step": 8327 }, { "epoch": 0.900032421917216, "grad_norm": 0.46582677960395813, "learning_rate": 8e-05, "loss": 1.7545, "step": 8328 }, { "epoch": 0.9001404949746028, "grad_norm": 0.44846415519714355, "learning_rate": 8e-05, "loss": 1.7553, "step": 8329 }, { "epoch": 0.9002485680319896, "grad_norm": 0.5072449445724487, "learning_rate": 8e-05, "loss": 1.765, "step": 8330 }, { "epoch": 0.9003566410893764, "grad_norm": 0.4355852007865906, "learning_rate": 8e-05, "loss": 1.7648, "step": 8331 }, { "epoch": 0.9004647141467632, "grad_norm": 0.4025591313838959, "learning_rate": 8e-05, "loss": 1.4684, "step": 8332 }, { "epoch": 0.90057278720415, "grad_norm": 0.4490521252155304, "learning_rate": 8e-05, "loss": 1.5859, "step": 8333 }, { "epoch": 0.9006808602615368, "grad_norm": 0.5083915591239929, "learning_rate": 8e-05, "loss": 1.8476, "step": 8334 }, { "epoch": 0.9007889333189236, "grad_norm": 0.4382005035877228, "learning_rate": 8e-05, "loss": 1.5957, "step": 8335 }, { "epoch": 0.9008970063763104, "grad_norm": 0.4660583734512329, "learning_rate": 8e-05, "loss": 1.7125, "step": 8336 }, { "epoch": 0.9010050794336972, "grad_norm": 0.47620731592178345, "learning_rate": 8e-05, "loss": 1.6999, "step": 8337 }, { "epoch": 0.9011131524910839, "grad_norm": 0.4604680836200714, "learning_rate": 8e-05, "loss": 1.5749, "step": 8338 }, { "epoch": 0.9012212255484707, "grad_norm": 0.43628013134002686, "learning_rate": 8e-05, "loss": 1.689, "step": 8339 }, { "epoch": 0.9013292986058575, "grad_norm": 0.4581226408481598, "learning_rate": 8e-05, "loss": 1.7103, "step": 8340 }, { "epoch": 0.9014373716632443, "grad_norm": 0.4472721517086029, "learning_rate": 8e-05, "loss": 1.7894, "step": 8341 }, { "epoch": 0.9015454447206311, "grad_norm": 0.45481887459754944, "learning_rate": 8e-05, "loss": 1.7743, "step": 8342 }, { "epoch": 0.9016535177780179, "grad_norm": 0.44724076986312866, "learning_rate": 8e-05, "loss": 1.7492, "step": 8343 }, { "epoch": 0.9017615908354047, "grad_norm": 0.4551074504852295, "learning_rate": 8e-05, "loss": 1.7118, "step": 8344 }, { "epoch": 0.9018696638927916, "grad_norm": 0.41832268238067627, "learning_rate": 8e-05, "loss": 1.8123, "step": 8345 }, { "epoch": 0.9019777369501784, "grad_norm": 0.5129278302192688, "learning_rate": 8e-05, "loss": 1.8673, "step": 8346 }, { "epoch": 0.9020858100075652, "grad_norm": 0.44523778557777405, "learning_rate": 8e-05, "loss": 1.742, "step": 8347 }, { "epoch": 0.902193883064952, "grad_norm": 0.42445966601371765, "learning_rate": 8e-05, "loss": 1.6472, "step": 8348 }, { "epoch": 0.9023019561223387, "grad_norm": 0.44649189710617065, "learning_rate": 8e-05, "loss": 1.6588, "step": 8349 }, { "epoch": 0.9024100291797255, "grad_norm": 0.41916000843048096, "learning_rate": 8e-05, "loss": 1.5735, "step": 8350 }, { "epoch": 0.9025181022371123, "grad_norm": 0.4576226770877838, "learning_rate": 8e-05, "loss": 1.7543, "step": 8351 }, { "epoch": 0.9026261752944991, "grad_norm": 0.43791699409484863, "learning_rate": 8e-05, "loss": 1.6419, "step": 8352 }, { "epoch": 0.9027342483518859, "grad_norm": 0.44870659708976746, "learning_rate": 8e-05, "loss": 1.6437, "step": 8353 }, { "epoch": 0.9028423214092727, "grad_norm": 0.46554985642433167, "learning_rate": 8e-05, "loss": 1.9304, "step": 8354 }, { "epoch": 0.9029503944666595, "grad_norm": 0.4765322804450989, "learning_rate": 8e-05, "loss": 1.7711, "step": 8355 }, { "epoch": 0.9030584675240463, "grad_norm": 0.44999149441719055, "learning_rate": 8e-05, "loss": 1.6841, "step": 8356 }, { "epoch": 0.9031665405814331, "grad_norm": 0.44177010655403137, "learning_rate": 8e-05, "loss": 1.7008, "step": 8357 }, { "epoch": 0.9032746136388199, "grad_norm": 0.4359041452407837, "learning_rate": 8e-05, "loss": 1.6078, "step": 8358 }, { "epoch": 0.9033826866962066, "grad_norm": 0.46943941712379456, "learning_rate": 8e-05, "loss": 1.6414, "step": 8359 }, { "epoch": 0.9034907597535934, "grad_norm": 0.5484488606452942, "learning_rate": 8e-05, "loss": 1.8352, "step": 8360 }, { "epoch": 0.9035988328109802, "grad_norm": 0.45179930329322815, "learning_rate": 8e-05, "loss": 1.4781, "step": 8361 }, { "epoch": 0.903706905868367, "grad_norm": 0.48167726397514343, "learning_rate": 8e-05, "loss": 1.6032, "step": 8362 }, { "epoch": 0.9038149789257538, "grad_norm": 0.481144517660141, "learning_rate": 8e-05, "loss": 1.737, "step": 8363 }, { "epoch": 0.9039230519831406, "grad_norm": 0.43257400393486023, "learning_rate": 8e-05, "loss": 1.6497, "step": 8364 }, { "epoch": 0.9040311250405274, "grad_norm": 0.48818886280059814, "learning_rate": 8e-05, "loss": 1.6743, "step": 8365 }, { "epoch": 0.9041391980979142, "grad_norm": 0.4478825032711029, "learning_rate": 8e-05, "loss": 1.5879, "step": 8366 }, { "epoch": 0.904247271155301, "grad_norm": 0.4254421293735504, "learning_rate": 8e-05, "loss": 1.5205, "step": 8367 }, { "epoch": 0.9043553442126878, "grad_norm": 0.5373061299324036, "learning_rate": 8e-05, "loss": 1.9563, "step": 8368 }, { "epoch": 0.9044634172700746, "grad_norm": 0.49477702379226685, "learning_rate": 8e-05, "loss": 1.6277, "step": 8369 }, { "epoch": 0.9045714903274613, "grad_norm": 0.4739716947078705, "learning_rate": 8e-05, "loss": 1.6967, "step": 8370 }, { "epoch": 0.9046795633848481, "grad_norm": 0.4908887445926666, "learning_rate": 8e-05, "loss": 1.8455, "step": 8371 }, { "epoch": 0.9047876364422349, "grad_norm": 0.44659674167633057, "learning_rate": 8e-05, "loss": 1.4958, "step": 8372 }, { "epoch": 0.9048957094996217, "grad_norm": 0.4344189465045929, "learning_rate": 8e-05, "loss": 1.7335, "step": 8373 }, { "epoch": 0.9050037825570085, "grad_norm": 0.4726675748825073, "learning_rate": 8e-05, "loss": 1.5947, "step": 8374 }, { "epoch": 0.9051118556143953, "grad_norm": 0.4412665367126465, "learning_rate": 8e-05, "loss": 1.6274, "step": 8375 }, { "epoch": 0.9052199286717821, "grad_norm": 0.4899728000164032, "learning_rate": 8e-05, "loss": 1.6558, "step": 8376 }, { "epoch": 0.9053280017291689, "grad_norm": 0.49524986743927, "learning_rate": 8e-05, "loss": 1.6514, "step": 8377 }, { "epoch": 0.9054360747865557, "grad_norm": 0.46752119064331055, "learning_rate": 8e-05, "loss": 1.8059, "step": 8378 }, { "epoch": 0.9055441478439425, "grad_norm": 0.4442811906337738, "learning_rate": 8e-05, "loss": 1.768, "step": 8379 }, { "epoch": 0.9056522209013294, "grad_norm": 0.4768247902393341, "learning_rate": 8e-05, "loss": 1.641, "step": 8380 }, { "epoch": 0.905760293958716, "grad_norm": 0.4237096905708313, "learning_rate": 8e-05, "loss": 1.6001, "step": 8381 }, { "epoch": 0.9058683670161028, "grad_norm": 0.5012621879577637, "learning_rate": 8e-05, "loss": 1.8321, "step": 8382 }, { "epoch": 0.9059764400734897, "grad_norm": 0.4492206275463104, "learning_rate": 8e-05, "loss": 1.5988, "step": 8383 }, { "epoch": 0.9060845131308765, "grad_norm": 0.42873579263687134, "learning_rate": 8e-05, "loss": 1.5753, "step": 8384 }, { "epoch": 0.9061925861882633, "grad_norm": 0.43720677495002747, "learning_rate": 8e-05, "loss": 1.5251, "step": 8385 }, { "epoch": 0.9063006592456501, "grad_norm": 0.43512728810310364, "learning_rate": 8e-05, "loss": 1.7606, "step": 8386 }, { "epoch": 0.9064087323030369, "grad_norm": 0.46988445520401, "learning_rate": 8e-05, "loss": 1.7972, "step": 8387 }, { "epoch": 0.9065168053604237, "grad_norm": 0.5579907298088074, "learning_rate": 8e-05, "loss": 1.4748, "step": 8388 }, { "epoch": 0.9066248784178105, "grad_norm": 0.4689333140850067, "learning_rate": 8e-05, "loss": 1.6548, "step": 8389 }, { "epoch": 0.9067329514751973, "grad_norm": 0.4443921446800232, "learning_rate": 8e-05, "loss": 1.4808, "step": 8390 }, { "epoch": 0.906841024532584, "grad_norm": 0.4699137210845947, "learning_rate": 8e-05, "loss": 1.6671, "step": 8391 }, { "epoch": 0.9069490975899708, "grad_norm": 0.4774796664714813, "learning_rate": 8e-05, "loss": 1.754, "step": 8392 }, { "epoch": 0.9070571706473576, "grad_norm": 0.46835875511169434, "learning_rate": 8e-05, "loss": 1.5621, "step": 8393 }, { "epoch": 0.9071652437047444, "grad_norm": 0.4202573597431183, "learning_rate": 8e-05, "loss": 1.4751, "step": 8394 }, { "epoch": 0.9072733167621312, "grad_norm": 0.4558355212211609, "learning_rate": 8e-05, "loss": 1.5842, "step": 8395 }, { "epoch": 0.907381389819518, "grad_norm": 0.48514053225517273, "learning_rate": 8e-05, "loss": 1.8309, "step": 8396 }, { "epoch": 0.9074894628769048, "grad_norm": 0.4573248624801636, "learning_rate": 8e-05, "loss": 1.7348, "step": 8397 }, { "epoch": 0.9075975359342916, "grad_norm": 0.41082918643951416, "learning_rate": 8e-05, "loss": 1.584, "step": 8398 }, { "epoch": 0.9077056089916784, "grad_norm": 0.44111937284469604, "learning_rate": 8e-05, "loss": 1.5298, "step": 8399 }, { "epoch": 0.9078136820490652, "grad_norm": 0.42922741174697876, "learning_rate": 8e-05, "loss": 1.6261, "step": 8400 }, { "epoch": 0.907921755106452, "grad_norm": 0.42778047919273376, "learning_rate": 8e-05, "loss": 1.6789, "step": 8401 }, { "epoch": 0.9080298281638387, "grad_norm": 0.40763190388679504, "learning_rate": 8e-05, "loss": 1.5788, "step": 8402 }, { "epoch": 0.9081379012212255, "grad_norm": 0.43524256348609924, "learning_rate": 8e-05, "loss": 1.584, "step": 8403 }, { "epoch": 0.9082459742786123, "grad_norm": 0.5226283669471741, "learning_rate": 8e-05, "loss": 1.8668, "step": 8404 }, { "epoch": 0.9083540473359991, "grad_norm": 0.48225507140159607, "learning_rate": 8e-05, "loss": 1.6829, "step": 8405 }, { "epoch": 0.9084621203933859, "grad_norm": 0.48837167024612427, "learning_rate": 8e-05, "loss": 1.7535, "step": 8406 }, { "epoch": 0.9085701934507727, "grad_norm": 0.45517635345458984, "learning_rate": 8e-05, "loss": 1.7603, "step": 8407 }, { "epoch": 0.9086782665081595, "grad_norm": 0.43195992708206177, "learning_rate": 8e-05, "loss": 1.6992, "step": 8408 }, { "epoch": 0.9087863395655463, "grad_norm": 0.43882372975349426, "learning_rate": 8e-05, "loss": 1.525, "step": 8409 }, { "epoch": 0.9088944126229331, "grad_norm": 0.5131354331970215, "learning_rate": 8e-05, "loss": 1.869, "step": 8410 }, { "epoch": 0.9090024856803199, "grad_norm": 0.4263278841972351, "learning_rate": 8e-05, "loss": 1.5243, "step": 8411 }, { "epoch": 0.9091105587377067, "grad_norm": 0.47755318880081177, "learning_rate": 8e-05, "loss": 1.9077, "step": 8412 }, { "epoch": 0.9092186317950934, "grad_norm": 0.44261181354522705, "learning_rate": 8e-05, "loss": 1.7442, "step": 8413 }, { "epoch": 0.9093267048524802, "grad_norm": 0.4749271869659424, "learning_rate": 8e-05, "loss": 1.8186, "step": 8414 }, { "epoch": 0.909434777909867, "grad_norm": 0.44900795817375183, "learning_rate": 8e-05, "loss": 1.7041, "step": 8415 }, { "epoch": 0.9095428509672538, "grad_norm": 0.4932079613208771, "learning_rate": 8e-05, "loss": 1.7726, "step": 8416 }, { "epoch": 0.9096509240246407, "grad_norm": 0.4421401917934418, "learning_rate": 8e-05, "loss": 1.6595, "step": 8417 }, { "epoch": 0.9097589970820275, "grad_norm": 0.4883693754673004, "learning_rate": 8e-05, "loss": 1.7818, "step": 8418 }, { "epoch": 0.9098670701394143, "grad_norm": 0.505214273929596, "learning_rate": 8e-05, "loss": 1.7465, "step": 8419 }, { "epoch": 0.9099751431968011, "grad_norm": 0.49120426177978516, "learning_rate": 8e-05, "loss": 1.7449, "step": 8420 }, { "epoch": 0.9100832162541879, "grad_norm": 0.4558044373989105, "learning_rate": 8e-05, "loss": 1.7221, "step": 8421 }, { "epoch": 0.9101912893115747, "grad_norm": 0.47297748923301697, "learning_rate": 8e-05, "loss": 1.7701, "step": 8422 }, { "epoch": 0.9102993623689615, "grad_norm": 0.43789345026016235, "learning_rate": 8e-05, "loss": 1.6474, "step": 8423 }, { "epoch": 0.9104074354263482, "grad_norm": 0.4335778057575226, "learning_rate": 8e-05, "loss": 1.5513, "step": 8424 }, { "epoch": 0.910515508483735, "grad_norm": 0.467765748500824, "learning_rate": 8e-05, "loss": 1.595, "step": 8425 }, { "epoch": 0.9106235815411218, "grad_norm": 0.439272403717041, "learning_rate": 8e-05, "loss": 1.6163, "step": 8426 }, { "epoch": 0.9107316545985086, "grad_norm": 0.4447386562824249, "learning_rate": 8e-05, "loss": 1.6776, "step": 8427 }, { "epoch": 0.9108397276558954, "grad_norm": 0.4384692907333374, "learning_rate": 8e-05, "loss": 1.7039, "step": 8428 }, { "epoch": 0.9109478007132822, "grad_norm": 0.43406757712364197, "learning_rate": 8e-05, "loss": 1.5657, "step": 8429 }, { "epoch": 0.911055873770669, "grad_norm": 0.43077048659324646, "learning_rate": 8e-05, "loss": 1.7525, "step": 8430 }, { "epoch": 0.9111639468280558, "grad_norm": 0.48872148990631104, "learning_rate": 8e-05, "loss": 1.7214, "step": 8431 }, { "epoch": 0.9112720198854426, "grad_norm": 0.4328186511993408, "learning_rate": 8e-05, "loss": 1.6992, "step": 8432 }, { "epoch": 0.9113800929428294, "grad_norm": 0.4932243824005127, "learning_rate": 8e-05, "loss": 1.8002, "step": 8433 }, { "epoch": 0.9114881660002161, "grad_norm": 0.4465275704860687, "learning_rate": 8e-05, "loss": 1.5523, "step": 8434 }, { "epoch": 0.9115962390576029, "grad_norm": 0.4065786600112915, "learning_rate": 8e-05, "loss": 1.4688, "step": 8435 }, { "epoch": 0.9117043121149897, "grad_norm": 0.5015559792518616, "learning_rate": 8e-05, "loss": 1.6494, "step": 8436 }, { "epoch": 0.9118123851723765, "grad_norm": 0.43104779720306396, "learning_rate": 8e-05, "loss": 1.6737, "step": 8437 }, { "epoch": 0.9119204582297633, "grad_norm": 0.4223703444004059, "learning_rate": 8e-05, "loss": 1.598, "step": 8438 }, { "epoch": 0.9120285312871501, "grad_norm": 0.43817225098609924, "learning_rate": 8e-05, "loss": 1.6462, "step": 8439 }, { "epoch": 0.9121366043445369, "grad_norm": 0.4483393430709839, "learning_rate": 8e-05, "loss": 1.7316, "step": 8440 }, { "epoch": 0.9122446774019237, "grad_norm": 0.4212431311607361, "learning_rate": 8e-05, "loss": 1.5841, "step": 8441 }, { "epoch": 0.9123527504593105, "grad_norm": 0.40971121191978455, "learning_rate": 8e-05, "loss": 1.6146, "step": 8442 }, { "epoch": 0.9124608235166973, "grad_norm": 0.4950661361217499, "learning_rate": 8e-05, "loss": 1.9215, "step": 8443 }, { "epoch": 0.9125688965740841, "grad_norm": 0.47274160385131836, "learning_rate": 8e-05, "loss": 1.7629, "step": 8444 }, { "epoch": 0.9126769696314708, "grad_norm": 0.4404848515987396, "learning_rate": 8e-05, "loss": 1.7703, "step": 8445 }, { "epoch": 0.9127850426888576, "grad_norm": 0.4422736167907715, "learning_rate": 8e-05, "loss": 1.6567, "step": 8446 }, { "epoch": 0.9128931157462444, "grad_norm": 0.46567273139953613, "learning_rate": 8e-05, "loss": 1.7431, "step": 8447 }, { "epoch": 0.9130011888036312, "grad_norm": 0.4471225142478943, "learning_rate": 8e-05, "loss": 1.74, "step": 8448 }, { "epoch": 0.913109261861018, "grad_norm": 0.44229182600975037, "learning_rate": 8e-05, "loss": 1.6931, "step": 8449 }, { "epoch": 0.9132173349184048, "grad_norm": 0.44149771332740784, "learning_rate": 8e-05, "loss": 1.6201, "step": 8450 }, { "epoch": 0.9133254079757916, "grad_norm": 0.49117380380630493, "learning_rate": 8e-05, "loss": 1.879, "step": 8451 }, { "epoch": 0.9134334810331785, "grad_norm": 0.4040817618370056, "learning_rate": 8e-05, "loss": 1.539, "step": 8452 }, { "epoch": 0.9135415540905653, "grad_norm": 0.43333080410957336, "learning_rate": 8e-05, "loss": 1.7306, "step": 8453 }, { "epoch": 0.9136496271479521, "grad_norm": 0.45971518754959106, "learning_rate": 8e-05, "loss": 1.3876, "step": 8454 }, { "epoch": 0.9137577002053389, "grad_norm": 0.4458027482032776, "learning_rate": 8e-05, "loss": 1.598, "step": 8455 }, { "epoch": 0.9138657732627256, "grad_norm": 0.4122157692909241, "learning_rate": 8e-05, "loss": 1.4831, "step": 8456 }, { "epoch": 0.9139738463201124, "grad_norm": 0.425391286611557, "learning_rate": 8e-05, "loss": 1.6019, "step": 8457 }, { "epoch": 0.9140819193774992, "grad_norm": 0.4485049545764923, "learning_rate": 8e-05, "loss": 1.6886, "step": 8458 }, { "epoch": 0.914189992434886, "grad_norm": 0.46797800064086914, "learning_rate": 8e-05, "loss": 1.7547, "step": 8459 }, { "epoch": 0.9142980654922728, "grad_norm": 0.44700464606285095, "learning_rate": 8e-05, "loss": 1.7615, "step": 8460 }, { "epoch": 0.9144061385496596, "grad_norm": 0.4325391948223114, "learning_rate": 8e-05, "loss": 1.5704, "step": 8461 }, { "epoch": 0.9145142116070464, "grad_norm": 0.48215997219085693, "learning_rate": 8e-05, "loss": 1.6378, "step": 8462 }, { "epoch": 0.9146222846644332, "grad_norm": 0.44423195719718933, "learning_rate": 8e-05, "loss": 1.7159, "step": 8463 }, { "epoch": 0.91473035772182, "grad_norm": 0.4147183299064636, "learning_rate": 8e-05, "loss": 1.6472, "step": 8464 }, { "epoch": 0.9148384307792068, "grad_norm": 0.45942768454551697, "learning_rate": 8e-05, "loss": 1.7815, "step": 8465 }, { "epoch": 0.9149465038365935, "grad_norm": 0.45035386085510254, "learning_rate": 8e-05, "loss": 1.7567, "step": 8466 }, { "epoch": 0.9150545768939803, "grad_norm": 0.5026825070381165, "learning_rate": 8e-05, "loss": 1.9804, "step": 8467 }, { "epoch": 0.9151626499513671, "grad_norm": 0.5101926922798157, "learning_rate": 8e-05, "loss": 1.5786, "step": 8468 }, { "epoch": 0.9152707230087539, "grad_norm": 0.43641337752342224, "learning_rate": 8e-05, "loss": 1.6791, "step": 8469 }, { "epoch": 0.9153787960661407, "grad_norm": 0.4372164309024811, "learning_rate": 8e-05, "loss": 1.7449, "step": 8470 }, { "epoch": 0.9154868691235275, "grad_norm": 0.4808759093284607, "learning_rate": 8e-05, "loss": 1.746, "step": 8471 }, { "epoch": 0.9155949421809143, "grad_norm": 0.437838613986969, "learning_rate": 8e-05, "loss": 1.7835, "step": 8472 }, { "epoch": 0.9157030152383011, "grad_norm": 0.41996559500694275, "learning_rate": 8e-05, "loss": 1.3469, "step": 8473 }, { "epoch": 0.9158110882956879, "grad_norm": 0.4352876842021942, "learning_rate": 8e-05, "loss": 1.6582, "step": 8474 }, { "epoch": 0.9159191613530747, "grad_norm": 0.4507425129413605, "learning_rate": 8e-05, "loss": 1.5614, "step": 8475 }, { "epoch": 0.9160272344104615, "grad_norm": 0.4445309638977051, "learning_rate": 8e-05, "loss": 1.5649, "step": 8476 }, { "epoch": 0.9161353074678482, "grad_norm": 0.5268515348434448, "learning_rate": 8e-05, "loss": 1.5517, "step": 8477 }, { "epoch": 0.916243380525235, "grad_norm": 0.4954555928707123, "learning_rate": 8e-05, "loss": 1.7447, "step": 8478 }, { "epoch": 0.9163514535826218, "grad_norm": 0.4643990099430084, "learning_rate": 8e-05, "loss": 1.7051, "step": 8479 }, { "epoch": 0.9164595266400086, "grad_norm": 0.49729764461517334, "learning_rate": 8e-05, "loss": 1.613, "step": 8480 }, { "epoch": 0.9165675996973954, "grad_norm": 0.45854613184928894, "learning_rate": 8e-05, "loss": 1.7485, "step": 8481 }, { "epoch": 0.9166756727547822, "grad_norm": 0.44590333104133606, "learning_rate": 8e-05, "loss": 1.5818, "step": 8482 }, { "epoch": 0.916783745812169, "grad_norm": 0.5046656131744385, "learning_rate": 8e-05, "loss": 1.8085, "step": 8483 }, { "epoch": 0.9168918188695558, "grad_norm": 0.4900107681751251, "learning_rate": 8e-05, "loss": 1.8337, "step": 8484 }, { "epoch": 0.9169998919269426, "grad_norm": 0.4426688551902771, "learning_rate": 8e-05, "loss": 1.5537, "step": 8485 }, { "epoch": 0.9171079649843294, "grad_norm": 0.44598665833473206, "learning_rate": 8e-05, "loss": 1.6782, "step": 8486 }, { "epoch": 0.9172160380417163, "grad_norm": 0.42534151673316956, "learning_rate": 8e-05, "loss": 1.7755, "step": 8487 }, { "epoch": 0.917324111099103, "grad_norm": 0.4556141495704651, "learning_rate": 8e-05, "loss": 1.5927, "step": 8488 }, { "epoch": 0.9174321841564897, "grad_norm": 0.47331130504608154, "learning_rate": 8e-05, "loss": 1.8379, "step": 8489 }, { "epoch": 0.9175402572138766, "grad_norm": 0.4154854118824005, "learning_rate": 8e-05, "loss": 1.6237, "step": 8490 }, { "epoch": 0.9176483302712634, "grad_norm": 0.43423759937286377, "learning_rate": 8e-05, "loss": 1.7523, "step": 8491 }, { "epoch": 0.9177564033286502, "grad_norm": 0.43821972608566284, "learning_rate": 8e-05, "loss": 1.658, "step": 8492 }, { "epoch": 0.917864476386037, "grad_norm": 0.44566819071769714, "learning_rate": 8e-05, "loss": 1.7668, "step": 8493 }, { "epoch": 0.9179725494434238, "grad_norm": 0.42382100224494934, "learning_rate": 8e-05, "loss": 1.7368, "step": 8494 }, { "epoch": 0.9180806225008106, "grad_norm": 0.47786596417427063, "learning_rate": 8e-05, "loss": 1.7978, "step": 8495 }, { "epoch": 0.9181886955581974, "grad_norm": 0.42632997035980225, "learning_rate": 8e-05, "loss": 1.6885, "step": 8496 }, { "epoch": 0.9182967686155842, "grad_norm": 0.4424653947353363, "learning_rate": 8e-05, "loss": 1.6913, "step": 8497 }, { "epoch": 0.9184048416729709, "grad_norm": 0.43649861216545105, "learning_rate": 8e-05, "loss": 1.5875, "step": 8498 }, { "epoch": 0.9185129147303577, "grad_norm": 0.46247681975364685, "learning_rate": 8e-05, "loss": 1.5597, "step": 8499 }, { "epoch": 0.9186209877877445, "grad_norm": 0.43868863582611084, "learning_rate": 8e-05, "loss": 1.6649, "step": 8500 }, { "epoch": 0.9187290608451313, "grad_norm": 0.46822860836982727, "learning_rate": 8e-05, "loss": 1.6099, "step": 8501 }, { "epoch": 0.9188371339025181, "grad_norm": 0.4385744631290436, "learning_rate": 8e-05, "loss": 1.6638, "step": 8502 }, { "epoch": 0.9189452069599049, "grad_norm": 0.45852941274642944, "learning_rate": 8e-05, "loss": 1.6195, "step": 8503 }, { "epoch": 0.9190532800172917, "grad_norm": 0.48376014828681946, "learning_rate": 8e-05, "loss": 1.8762, "step": 8504 }, { "epoch": 0.9191613530746785, "grad_norm": 0.44886714220046997, "learning_rate": 8e-05, "loss": 1.7561, "step": 8505 }, { "epoch": 0.9192694261320653, "grad_norm": 0.4467584192752838, "learning_rate": 8e-05, "loss": 1.7236, "step": 8506 }, { "epoch": 0.9193774991894521, "grad_norm": 0.47801363468170166, "learning_rate": 8e-05, "loss": 1.7194, "step": 8507 }, { "epoch": 0.9194855722468389, "grad_norm": 0.5006362795829773, "learning_rate": 8e-05, "loss": 1.8219, "step": 8508 }, { "epoch": 0.9195936453042256, "grad_norm": 0.4319107234477997, "learning_rate": 8e-05, "loss": 1.6312, "step": 8509 }, { "epoch": 0.9197017183616124, "grad_norm": 0.49014168977737427, "learning_rate": 8e-05, "loss": 1.8997, "step": 8510 }, { "epoch": 0.9198097914189992, "grad_norm": 0.42923304438591003, "learning_rate": 8e-05, "loss": 1.6663, "step": 8511 }, { "epoch": 0.919917864476386, "grad_norm": 0.44929423928260803, "learning_rate": 8e-05, "loss": 1.7784, "step": 8512 }, { "epoch": 0.9200259375337728, "grad_norm": 0.46725496649742126, "learning_rate": 8e-05, "loss": 1.6346, "step": 8513 }, { "epoch": 0.9201340105911596, "grad_norm": 0.4516736567020416, "learning_rate": 8e-05, "loss": 1.7014, "step": 8514 }, { "epoch": 0.9202420836485464, "grad_norm": 0.48717013001441956, "learning_rate": 8e-05, "loss": 1.8207, "step": 8515 }, { "epoch": 0.9203501567059332, "grad_norm": 0.46576979756355286, "learning_rate": 8e-05, "loss": 1.6679, "step": 8516 }, { "epoch": 0.92045822976332, "grad_norm": 0.473868727684021, "learning_rate": 8e-05, "loss": 1.607, "step": 8517 }, { "epoch": 0.9205663028207068, "grad_norm": 0.4697057604789734, "learning_rate": 8e-05, "loss": 1.6608, "step": 8518 }, { "epoch": 0.9206743758780936, "grad_norm": 0.47592538595199585, "learning_rate": 8e-05, "loss": 1.7659, "step": 8519 }, { "epoch": 0.9207824489354803, "grad_norm": 0.46058976650238037, "learning_rate": 8e-05, "loss": 1.6524, "step": 8520 }, { "epoch": 0.9208905219928671, "grad_norm": 0.5086554884910583, "learning_rate": 8e-05, "loss": 1.8823, "step": 8521 }, { "epoch": 0.9209985950502539, "grad_norm": 0.47187140583992004, "learning_rate": 8e-05, "loss": 1.9168, "step": 8522 }, { "epoch": 0.9211066681076407, "grad_norm": 0.4969559907913208, "learning_rate": 8e-05, "loss": 1.703, "step": 8523 }, { "epoch": 0.9212147411650276, "grad_norm": 0.503826916217804, "learning_rate": 8e-05, "loss": 1.9277, "step": 8524 }, { "epoch": 0.9213228142224144, "grad_norm": 0.42533379793167114, "learning_rate": 8e-05, "loss": 1.6742, "step": 8525 }, { "epoch": 0.9214308872798012, "grad_norm": 0.44017961621284485, "learning_rate": 8e-05, "loss": 1.6435, "step": 8526 }, { "epoch": 0.921538960337188, "grad_norm": 0.48178631067276, "learning_rate": 8e-05, "loss": 1.5879, "step": 8527 }, { "epoch": 0.9216470333945748, "grad_norm": 0.4938312768936157, "learning_rate": 8e-05, "loss": 1.7164, "step": 8528 }, { "epoch": 0.9217551064519616, "grad_norm": 0.45259562134742737, "learning_rate": 8e-05, "loss": 1.6236, "step": 8529 }, { "epoch": 0.9218631795093484, "grad_norm": 0.4626030921936035, "learning_rate": 8e-05, "loss": 1.7506, "step": 8530 }, { "epoch": 0.9219712525667351, "grad_norm": 0.4296257793903351, "learning_rate": 8e-05, "loss": 1.6074, "step": 8531 }, { "epoch": 0.9220793256241219, "grad_norm": 0.4375074803829193, "learning_rate": 8e-05, "loss": 1.5612, "step": 8532 }, { "epoch": 0.9221873986815087, "grad_norm": 0.4593726396560669, "learning_rate": 8e-05, "loss": 1.6899, "step": 8533 }, { "epoch": 0.9222954717388955, "grad_norm": 0.43427368998527527, "learning_rate": 8e-05, "loss": 1.4534, "step": 8534 }, { "epoch": 0.9224035447962823, "grad_norm": 0.503317654132843, "learning_rate": 8e-05, "loss": 1.6894, "step": 8535 }, { "epoch": 0.9225116178536691, "grad_norm": 0.46713751554489136, "learning_rate": 8e-05, "loss": 1.6931, "step": 8536 }, { "epoch": 0.9226196909110559, "grad_norm": 0.44700583815574646, "learning_rate": 8e-05, "loss": 1.7547, "step": 8537 }, { "epoch": 0.9227277639684427, "grad_norm": 0.4928402602672577, "learning_rate": 8e-05, "loss": 1.7227, "step": 8538 }, { "epoch": 0.9228358370258295, "grad_norm": 0.4513237774372101, "learning_rate": 8e-05, "loss": 1.6305, "step": 8539 }, { "epoch": 0.9229439100832163, "grad_norm": 0.41738131642341614, "learning_rate": 8e-05, "loss": 1.652, "step": 8540 }, { "epoch": 0.923051983140603, "grad_norm": 0.5020430684089661, "learning_rate": 8e-05, "loss": 1.8624, "step": 8541 }, { "epoch": 0.9231600561979898, "grad_norm": 0.45778968930244446, "learning_rate": 8e-05, "loss": 1.7325, "step": 8542 }, { "epoch": 0.9232681292553766, "grad_norm": 0.4430947005748749, "learning_rate": 8e-05, "loss": 1.6233, "step": 8543 }, { "epoch": 0.9233762023127634, "grad_norm": 0.4086894094944, "learning_rate": 8e-05, "loss": 1.5803, "step": 8544 }, { "epoch": 0.9234842753701502, "grad_norm": 0.4580029249191284, "learning_rate": 8e-05, "loss": 1.8953, "step": 8545 }, { "epoch": 0.923592348427537, "grad_norm": 0.44282546639442444, "learning_rate": 8e-05, "loss": 1.643, "step": 8546 }, { "epoch": 0.9237004214849238, "grad_norm": 0.4454610049724579, "learning_rate": 8e-05, "loss": 1.5715, "step": 8547 }, { "epoch": 0.9238084945423106, "grad_norm": 0.45537588000297546, "learning_rate": 8e-05, "loss": 1.5605, "step": 8548 }, { "epoch": 0.9239165675996974, "grad_norm": 0.42520013451576233, "learning_rate": 8e-05, "loss": 1.6757, "step": 8549 }, { "epoch": 0.9240246406570842, "grad_norm": 0.4468042552471161, "learning_rate": 8e-05, "loss": 1.6287, "step": 8550 }, { "epoch": 0.924132713714471, "grad_norm": 0.43673446774482727, "learning_rate": 8e-05, "loss": 1.6091, "step": 8551 }, { "epoch": 0.9242407867718577, "grad_norm": 0.451103150844574, "learning_rate": 8e-05, "loss": 1.6135, "step": 8552 }, { "epoch": 0.9243488598292445, "grad_norm": 0.45302942395210266, "learning_rate": 8e-05, "loss": 1.7488, "step": 8553 }, { "epoch": 0.9244569328866313, "grad_norm": 0.4815030097961426, "learning_rate": 8e-05, "loss": 1.7465, "step": 8554 }, { "epoch": 0.9245650059440181, "grad_norm": 0.42399755120277405, "learning_rate": 8e-05, "loss": 1.6228, "step": 8555 }, { "epoch": 0.9246730790014049, "grad_norm": 0.4627233147621155, "learning_rate": 8e-05, "loss": 1.6103, "step": 8556 }, { "epoch": 0.9247811520587917, "grad_norm": 0.44619399309158325, "learning_rate": 8e-05, "loss": 1.4978, "step": 8557 }, { "epoch": 0.9248892251161785, "grad_norm": 0.42383649945259094, "learning_rate": 8e-05, "loss": 1.6103, "step": 8558 }, { "epoch": 0.9249972981735654, "grad_norm": 0.4760134518146515, "learning_rate": 8e-05, "loss": 1.6343, "step": 8559 }, { "epoch": 0.9251053712309522, "grad_norm": 0.4912257492542267, "learning_rate": 8e-05, "loss": 1.7981, "step": 8560 }, { "epoch": 0.925213444288339, "grad_norm": 0.45822444558143616, "learning_rate": 8e-05, "loss": 1.6153, "step": 8561 }, { "epoch": 0.9253215173457258, "grad_norm": 0.45090505480766296, "learning_rate": 8e-05, "loss": 1.5721, "step": 8562 }, { "epoch": 0.9254295904031125, "grad_norm": 0.47562095522880554, "learning_rate": 8e-05, "loss": 1.7154, "step": 8563 }, { "epoch": 0.9255376634604993, "grad_norm": 0.4245311915874481, "learning_rate": 8e-05, "loss": 1.6615, "step": 8564 }, { "epoch": 0.9256457365178861, "grad_norm": 0.4544941186904907, "learning_rate": 8e-05, "loss": 1.6169, "step": 8565 }, { "epoch": 0.9257538095752729, "grad_norm": 0.44698649644851685, "learning_rate": 8e-05, "loss": 1.6859, "step": 8566 }, { "epoch": 0.9258618826326597, "grad_norm": 0.5071626305580139, "learning_rate": 8e-05, "loss": 1.7168, "step": 8567 }, { "epoch": 0.9259699556900465, "grad_norm": 0.45645830035209656, "learning_rate": 8e-05, "loss": 1.5716, "step": 8568 }, { "epoch": 0.9260780287474333, "grad_norm": 0.4296063184738159, "learning_rate": 8e-05, "loss": 1.5401, "step": 8569 }, { "epoch": 0.9261861018048201, "grad_norm": 0.4444999098777771, "learning_rate": 8e-05, "loss": 1.7043, "step": 8570 }, { "epoch": 0.9262941748622069, "grad_norm": 0.44081345200538635, "learning_rate": 8e-05, "loss": 1.6071, "step": 8571 }, { "epoch": 0.9264022479195937, "grad_norm": 0.42581984400749207, "learning_rate": 8e-05, "loss": 1.5478, "step": 8572 }, { "epoch": 0.9265103209769804, "grad_norm": 0.4816378057003021, "learning_rate": 8e-05, "loss": 1.7439, "step": 8573 }, { "epoch": 0.9266183940343672, "grad_norm": 0.4536915421485901, "learning_rate": 8e-05, "loss": 1.6375, "step": 8574 }, { "epoch": 0.926726467091754, "grad_norm": 0.4531098008155823, "learning_rate": 8e-05, "loss": 1.777, "step": 8575 }, { "epoch": 0.9268345401491408, "grad_norm": 0.47364217042922974, "learning_rate": 8e-05, "loss": 1.7583, "step": 8576 }, { "epoch": 0.9269426132065276, "grad_norm": 0.49039024114608765, "learning_rate": 8e-05, "loss": 1.6852, "step": 8577 }, { "epoch": 0.9270506862639144, "grad_norm": 0.4952732026576996, "learning_rate": 8e-05, "loss": 1.637, "step": 8578 }, { "epoch": 0.9271587593213012, "grad_norm": 0.43139296770095825, "learning_rate": 8e-05, "loss": 1.5272, "step": 8579 }, { "epoch": 0.927266832378688, "grad_norm": 0.40682005882263184, "learning_rate": 8e-05, "loss": 1.5532, "step": 8580 }, { "epoch": 0.9273749054360748, "grad_norm": 0.431856244802475, "learning_rate": 8e-05, "loss": 1.6411, "step": 8581 }, { "epoch": 0.9274829784934616, "grad_norm": 0.4694143235683441, "learning_rate": 8e-05, "loss": 1.7572, "step": 8582 }, { "epoch": 0.9275910515508484, "grad_norm": 0.4497951567173004, "learning_rate": 8e-05, "loss": 1.6207, "step": 8583 }, { "epoch": 0.9276991246082351, "grad_norm": 0.4084416627883911, "learning_rate": 8e-05, "loss": 1.487, "step": 8584 }, { "epoch": 0.9278071976656219, "grad_norm": 0.42258840799331665, "learning_rate": 8e-05, "loss": 1.5944, "step": 8585 }, { "epoch": 0.9279152707230087, "grad_norm": 0.46630358695983887, "learning_rate": 8e-05, "loss": 1.6223, "step": 8586 }, { "epoch": 0.9280233437803955, "grad_norm": 0.44153153896331787, "learning_rate": 8e-05, "loss": 1.6971, "step": 8587 }, { "epoch": 0.9281314168377823, "grad_norm": 0.5233045816421509, "learning_rate": 8e-05, "loss": 1.8136, "step": 8588 }, { "epoch": 0.9282394898951691, "grad_norm": 0.42860323190689087, "learning_rate": 8e-05, "loss": 1.5097, "step": 8589 }, { "epoch": 0.9283475629525559, "grad_norm": 0.46287471055984497, "learning_rate": 8e-05, "loss": 1.6938, "step": 8590 }, { "epoch": 0.9284556360099427, "grad_norm": 0.42428621649742126, "learning_rate": 8e-05, "loss": 1.566, "step": 8591 }, { "epoch": 0.9285637090673295, "grad_norm": 0.471026748418808, "learning_rate": 8e-05, "loss": 1.588, "step": 8592 }, { "epoch": 0.9286717821247163, "grad_norm": 0.458490788936615, "learning_rate": 8e-05, "loss": 1.6263, "step": 8593 }, { "epoch": 0.9287798551821032, "grad_norm": 0.42364415526390076, "learning_rate": 8e-05, "loss": 1.5462, "step": 8594 }, { "epoch": 0.9288879282394898, "grad_norm": 0.4703303575515747, "learning_rate": 8e-05, "loss": 1.7046, "step": 8595 }, { "epoch": 0.9289960012968767, "grad_norm": 0.4652478098869324, "learning_rate": 8e-05, "loss": 1.8245, "step": 8596 }, { "epoch": 0.9291040743542635, "grad_norm": 0.47196632623672485, "learning_rate": 8e-05, "loss": 1.6691, "step": 8597 }, { "epoch": 0.9292121474116503, "grad_norm": 0.4999677538871765, "learning_rate": 8e-05, "loss": 1.6947, "step": 8598 }, { "epoch": 0.9293202204690371, "grad_norm": 0.4324856102466583, "learning_rate": 8e-05, "loss": 1.6174, "step": 8599 }, { "epoch": 0.9294282935264239, "grad_norm": 0.4573093056678772, "learning_rate": 8e-05, "loss": 1.7805, "step": 8600 }, { "epoch": 0.9295363665838107, "grad_norm": 0.43886828422546387, "learning_rate": 8e-05, "loss": 1.6702, "step": 8601 }, { "epoch": 0.9296444396411975, "grad_norm": 0.4867112636566162, "learning_rate": 8e-05, "loss": 1.8293, "step": 8602 }, { "epoch": 0.9297525126985843, "grad_norm": 0.43762078881263733, "learning_rate": 8e-05, "loss": 1.6115, "step": 8603 }, { "epoch": 0.9298605857559711, "grad_norm": 0.48250192403793335, "learning_rate": 8e-05, "loss": 1.7169, "step": 8604 }, { "epoch": 0.9299686588133579, "grad_norm": 0.49352604150772095, "learning_rate": 8e-05, "loss": 1.8628, "step": 8605 }, { "epoch": 0.9300767318707446, "grad_norm": 0.44271957874298096, "learning_rate": 8e-05, "loss": 1.6553, "step": 8606 }, { "epoch": 0.9301848049281314, "grad_norm": 0.44248366355895996, "learning_rate": 8e-05, "loss": 1.6837, "step": 8607 }, { "epoch": 0.9302928779855182, "grad_norm": 0.5202497243881226, "learning_rate": 8e-05, "loss": 1.6768, "step": 8608 }, { "epoch": 0.930400951042905, "grad_norm": 0.4563630223274231, "learning_rate": 8e-05, "loss": 1.5228, "step": 8609 }, { "epoch": 0.9305090241002918, "grad_norm": 0.43811869621276855, "learning_rate": 8e-05, "loss": 1.7453, "step": 8610 }, { "epoch": 0.9306170971576786, "grad_norm": 0.4099179804325104, "learning_rate": 8e-05, "loss": 1.671, "step": 8611 }, { "epoch": 0.9307251702150654, "grad_norm": 0.4291231036186218, "learning_rate": 8e-05, "loss": 1.7304, "step": 8612 }, { "epoch": 0.9308332432724522, "grad_norm": 0.42774325609207153, "learning_rate": 8e-05, "loss": 1.5602, "step": 8613 }, { "epoch": 0.930941316329839, "grad_norm": 0.41870468854904175, "learning_rate": 8e-05, "loss": 1.6061, "step": 8614 }, { "epoch": 0.9310493893872258, "grad_norm": 0.4208846688270569, "learning_rate": 8e-05, "loss": 1.6162, "step": 8615 }, { "epoch": 0.9311574624446125, "grad_norm": 0.40778690576553345, "learning_rate": 8e-05, "loss": 1.4394, "step": 8616 }, { "epoch": 0.9312655355019993, "grad_norm": 0.4649544954299927, "learning_rate": 8e-05, "loss": 1.6239, "step": 8617 }, { "epoch": 0.9313736085593861, "grad_norm": 0.47690925002098083, "learning_rate": 8e-05, "loss": 1.7876, "step": 8618 }, { "epoch": 0.9314816816167729, "grad_norm": 0.44029635190963745, "learning_rate": 8e-05, "loss": 1.6682, "step": 8619 }, { "epoch": 0.9315897546741597, "grad_norm": 0.4435602128505707, "learning_rate": 8e-05, "loss": 1.8031, "step": 8620 }, { "epoch": 0.9316978277315465, "grad_norm": 0.49218258261680603, "learning_rate": 8e-05, "loss": 1.6774, "step": 8621 }, { "epoch": 0.9318059007889333, "grad_norm": 0.4245759844779968, "learning_rate": 8e-05, "loss": 1.6373, "step": 8622 }, { "epoch": 0.9319139738463201, "grad_norm": 0.4570235013961792, "learning_rate": 8e-05, "loss": 1.4736, "step": 8623 }, { "epoch": 0.9320220469037069, "grad_norm": 0.4461497962474823, "learning_rate": 8e-05, "loss": 1.638, "step": 8624 }, { "epoch": 0.9321301199610937, "grad_norm": 0.4578656256198883, "learning_rate": 8e-05, "loss": 1.7672, "step": 8625 }, { "epoch": 0.9322381930184805, "grad_norm": 0.41281411051750183, "learning_rate": 8e-05, "loss": 1.4554, "step": 8626 }, { "epoch": 0.9323462660758672, "grad_norm": 0.4288429915904999, "learning_rate": 8e-05, "loss": 1.6644, "step": 8627 }, { "epoch": 0.932454339133254, "grad_norm": 0.42713144421577454, "learning_rate": 8e-05, "loss": 1.6055, "step": 8628 }, { "epoch": 0.9325624121906408, "grad_norm": 0.4658537209033966, "learning_rate": 8e-05, "loss": 1.8696, "step": 8629 }, { "epoch": 0.9326704852480276, "grad_norm": 0.4469067454338074, "learning_rate": 8e-05, "loss": 1.5112, "step": 8630 }, { "epoch": 0.9327785583054145, "grad_norm": 0.43879175186157227, "learning_rate": 8e-05, "loss": 1.7084, "step": 8631 }, { "epoch": 0.9328866313628013, "grad_norm": 0.4571855366230011, "learning_rate": 8e-05, "loss": 1.6467, "step": 8632 }, { "epoch": 0.9329947044201881, "grad_norm": 0.4307180345058441, "learning_rate": 8e-05, "loss": 1.6392, "step": 8633 }, { "epoch": 0.9331027774775749, "grad_norm": 0.5130059719085693, "learning_rate": 8e-05, "loss": 1.8635, "step": 8634 }, { "epoch": 0.9332108505349617, "grad_norm": 0.4366745352745056, "learning_rate": 8e-05, "loss": 1.6928, "step": 8635 }, { "epoch": 0.9333189235923485, "grad_norm": 0.45282596349716187, "learning_rate": 8e-05, "loss": 1.5973, "step": 8636 }, { "epoch": 0.9334269966497353, "grad_norm": 0.4530145227909088, "learning_rate": 8e-05, "loss": 1.6461, "step": 8637 }, { "epoch": 0.933535069707122, "grad_norm": 0.45492392778396606, "learning_rate": 8e-05, "loss": 1.7291, "step": 8638 }, { "epoch": 0.9336431427645088, "grad_norm": 0.4647105038166046, "learning_rate": 8e-05, "loss": 1.5882, "step": 8639 }, { "epoch": 0.9337512158218956, "grad_norm": 0.4733675420284271, "learning_rate": 8e-05, "loss": 1.65, "step": 8640 }, { "epoch": 0.9338592888792824, "grad_norm": 0.4559079110622406, "learning_rate": 8e-05, "loss": 1.6811, "step": 8641 }, { "epoch": 0.9339673619366692, "grad_norm": 0.44133222103118896, "learning_rate": 8e-05, "loss": 1.5876, "step": 8642 }, { "epoch": 0.934075434994056, "grad_norm": 0.4527602195739746, "learning_rate": 8e-05, "loss": 1.6741, "step": 8643 }, { "epoch": 0.9341835080514428, "grad_norm": 0.4137949049472809, "learning_rate": 8e-05, "loss": 1.6126, "step": 8644 }, { "epoch": 0.9342915811088296, "grad_norm": 0.4451518654823303, "learning_rate": 8e-05, "loss": 1.8419, "step": 8645 }, { "epoch": 0.9343996541662164, "grad_norm": 0.4969654381275177, "learning_rate": 8e-05, "loss": 1.8215, "step": 8646 }, { "epoch": 0.9345077272236032, "grad_norm": 0.43897080421447754, "learning_rate": 8e-05, "loss": 1.6489, "step": 8647 }, { "epoch": 0.9346158002809899, "grad_norm": 0.4209134578704834, "learning_rate": 8e-05, "loss": 1.5378, "step": 8648 }, { "epoch": 0.9347238733383767, "grad_norm": 0.4578409492969513, "learning_rate": 8e-05, "loss": 1.5894, "step": 8649 }, { "epoch": 0.9348319463957635, "grad_norm": 0.44113144278526306, "learning_rate": 8e-05, "loss": 1.5539, "step": 8650 }, { "epoch": 0.9349400194531503, "grad_norm": 0.4748457968235016, "learning_rate": 8e-05, "loss": 1.6589, "step": 8651 }, { "epoch": 0.9350480925105371, "grad_norm": 0.47231242060661316, "learning_rate": 8e-05, "loss": 1.6749, "step": 8652 }, { "epoch": 0.9351561655679239, "grad_norm": 0.48061051964759827, "learning_rate": 8e-05, "loss": 1.6958, "step": 8653 }, { "epoch": 0.9352642386253107, "grad_norm": 0.517837405204773, "learning_rate": 8e-05, "loss": 1.829, "step": 8654 }, { "epoch": 0.9353723116826975, "grad_norm": 0.44026702642440796, "learning_rate": 8e-05, "loss": 1.4801, "step": 8655 }, { "epoch": 0.9354803847400843, "grad_norm": 0.42673078179359436, "learning_rate": 8e-05, "loss": 1.6031, "step": 8656 }, { "epoch": 0.9355884577974711, "grad_norm": 0.4543079733848572, "learning_rate": 8e-05, "loss": 1.6339, "step": 8657 }, { "epoch": 0.9356965308548579, "grad_norm": 0.4202789068222046, "learning_rate": 8e-05, "loss": 1.6015, "step": 8658 }, { "epoch": 0.9358046039122446, "grad_norm": 0.48994186520576477, "learning_rate": 8e-05, "loss": 1.8481, "step": 8659 }, { "epoch": 0.9359126769696314, "grad_norm": 0.4988189935684204, "learning_rate": 8e-05, "loss": 1.5841, "step": 8660 }, { "epoch": 0.9360207500270182, "grad_norm": 0.5137591361999512, "learning_rate": 8e-05, "loss": 1.7917, "step": 8661 }, { "epoch": 0.936128823084405, "grad_norm": 0.5027785301208496, "learning_rate": 8e-05, "loss": 1.772, "step": 8662 }, { "epoch": 0.9362368961417918, "grad_norm": 0.46774330735206604, "learning_rate": 8e-05, "loss": 1.6234, "step": 8663 }, { "epoch": 0.9363449691991786, "grad_norm": 0.45375704765319824, "learning_rate": 8e-05, "loss": 1.6513, "step": 8664 }, { "epoch": 0.9364530422565654, "grad_norm": 0.457907497882843, "learning_rate": 8e-05, "loss": 1.6248, "step": 8665 }, { "epoch": 0.9365611153139523, "grad_norm": 0.4319648742675781, "learning_rate": 8e-05, "loss": 1.7093, "step": 8666 }, { "epoch": 0.9366691883713391, "grad_norm": 0.4224325716495514, "learning_rate": 8e-05, "loss": 1.4312, "step": 8667 }, { "epoch": 0.9367772614287259, "grad_norm": 0.4418303966522217, "learning_rate": 8e-05, "loss": 1.7044, "step": 8668 }, { "epoch": 0.9368853344861127, "grad_norm": 0.4134460687637329, "learning_rate": 8e-05, "loss": 1.4708, "step": 8669 }, { "epoch": 0.9369934075434994, "grad_norm": 0.40765270590782166, "learning_rate": 8e-05, "loss": 1.5828, "step": 8670 }, { "epoch": 0.9371014806008862, "grad_norm": 0.4863780736923218, "learning_rate": 8e-05, "loss": 1.6492, "step": 8671 }, { "epoch": 0.937209553658273, "grad_norm": 0.4267246127128601, "learning_rate": 8e-05, "loss": 1.5651, "step": 8672 }, { "epoch": 0.9373176267156598, "grad_norm": 0.4770742356777191, "learning_rate": 8e-05, "loss": 1.6648, "step": 8673 }, { "epoch": 0.9374256997730466, "grad_norm": 0.4789092540740967, "learning_rate": 8e-05, "loss": 1.7472, "step": 8674 }, { "epoch": 0.9375337728304334, "grad_norm": 0.4715554714202881, "learning_rate": 8e-05, "loss": 1.7238, "step": 8675 }, { "epoch": 0.9376418458878202, "grad_norm": 0.4455958604812622, "learning_rate": 8e-05, "loss": 1.6389, "step": 8676 }, { "epoch": 0.937749918945207, "grad_norm": 0.49763575196266174, "learning_rate": 8e-05, "loss": 1.7187, "step": 8677 }, { "epoch": 0.9378579920025938, "grad_norm": 0.4251396656036377, "learning_rate": 8e-05, "loss": 1.5664, "step": 8678 }, { "epoch": 0.9379660650599806, "grad_norm": 0.4517434537410736, "learning_rate": 8e-05, "loss": 1.6265, "step": 8679 }, { "epoch": 0.9380741381173674, "grad_norm": 0.4603738486766815, "learning_rate": 8e-05, "loss": 1.7654, "step": 8680 }, { "epoch": 0.9381822111747541, "grad_norm": 0.4487018883228302, "learning_rate": 8e-05, "loss": 1.6441, "step": 8681 }, { "epoch": 0.9382902842321409, "grad_norm": 0.4316936433315277, "learning_rate": 8e-05, "loss": 1.6469, "step": 8682 }, { "epoch": 0.9383983572895277, "grad_norm": 0.5112162828445435, "learning_rate": 8e-05, "loss": 1.824, "step": 8683 }, { "epoch": 0.9385064303469145, "grad_norm": 0.46487748622894287, "learning_rate": 8e-05, "loss": 1.6663, "step": 8684 }, { "epoch": 0.9386145034043013, "grad_norm": 0.42839497327804565, "learning_rate": 8e-05, "loss": 1.5403, "step": 8685 }, { "epoch": 0.9387225764616881, "grad_norm": 0.4405612051486969, "learning_rate": 8e-05, "loss": 1.6271, "step": 8686 }, { "epoch": 0.9388306495190749, "grad_norm": 0.4270807206630707, "learning_rate": 8e-05, "loss": 1.5873, "step": 8687 }, { "epoch": 0.9389387225764617, "grad_norm": 0.4621935188770294, "learning_rate": 8e-05, "loss": 1.875, "step": 8688 }, { "epoch": 0.9390467956338485, "grad_norm": 0.449214369058609, "learning_rate": 8e-05, "loss": 1.6804, "step": 8689 }, { "epoch": 0.9391548686912353, "grad_norm": 0.46026885509490967, "learning_rate": 8e-05, "loss": 1.7182, "step": 8690 }, { "epoch": 0.939262941748622, "grad_norm": 0.42823758721351624, "learning_rate": 8e-05, "loss": 1.6124, "step": 8691 }, { "epoch": 0.9393710148060088, "grad_norm": 0.43850818276405334, "learning_rate": 8e-05, "loss": 1.6697, "step": 8692 }, { "epoch": 0.9394790878633956, "grad_norm": 0.4436233937740326, "learning_rate": 8e-05, "loss": 1.6622, "step": 8693 }, { "epoch": 0.9395871609207824, "grad_norm": 0.4780961573123932, "learning_rate": 8e-05, "loss": 1.8786, "step": 8694 }, { "epoch": 0.9396952339781692, "grad_norm": 0.45212769508361816, "learning_rate": 8e-05, "loss": 1.6566, "step": 8695 }, { "epoch": 0.939803307035556, "grad_norm": 0.4418102502822876, "learning_rate": 8e-05, "loss": 1.7551, "step": 8696 }, { "epoch": 0.9399113800929428, "grad_norm": 0.44285711646080017, "learning_rate": 8e-05, "loss": 1.6815, "step": 8697 }, { "epoch": 0.9400194531503296, "grad_norm": 0.4786936342716217, "learning_rate": 8e-05, "loss": 1.7087, "step": 8698 }, { "epoch": 0.9401275262077164, "grad_norm": 0.42566215991973877, "learning_rate": 8e-05, "loss": 1.6733, "step": 8699 }, { "epoch": 0.9402355992651033, "grad_norm": 0.45848917961120605, "learning_rate": 8e-05, "loss": 1.7491, "step": 8700 }, { "epoch": 0.94034367232249, "grad_norm": 0.5410504341125488, "learning_rate": 8e-05, "loss": 1.6164, "step": 8701 }, { "epoch": 0.9404517453798767, "grad_norm": 0.43278977274894714, "learning_rate": 8e-05, "loss": 1.6991, "step": 8702 }, { "epoch": 0.9405598184372636, "grad_norm": 0.4753572940826416, "learning_rate": 8e-05, "loss": 1.6511, "step": 8703 }, { "epoch": 0.9406678914946504, "grad_norm": 0.462433785200119, "learning_rate": 8e-05, "loss": 1.605, "step": 8704 }, { "epoch": 0.9407759645520372, "grad_norm": 0.47761476039886475, "learning_rate": 8e-05, "loss": 1.9282, "step": 8705 }, { "epoch": 0.940884037609424, "grad_norm": 0.4349345266819, "learning_rate": 8e-05, "loss": 1.592, "step": 8706 }, { "epoch": 0.9409921106668108, "grad_norm": 0.47425273060798645, "learning_rate": 8e-05, "loss": 1.7236, "step": 8707 }, { "epoch": 0.9411001837241976, "grad_norm": 0.4499209225177765, "learning_rate": 8e-05, "loss": 1.6556, "step": 8708 }, { "epoch": 0.9412082567815844, "grad_norm": 0.46320411562919617, "learning_rate": 8e-05, "loss": 1.8142, "step": 8709 }, { "epoch": 0.9413163298389712, "grad_norm": 0.4260393977165222, "learning_rate": 8e-05, "loss": 1.5865, "step": 8710 }, { "epoch": 0.941424402896358, "grad_norm": 0.44664281606674194, "learning_rate": 8e-05, "loss": 1.6188, "step": 8711 }, { "epoch": 0.9415324759537448, "grad_norm": 0.4603080749511719, "learning_rate": 8e-05, "loss": 1.8071, "step": 8712 }, { "epoch": 0.9416405490111315, "grad_norm": 0.5036304593086243, "learning_rate": 8e-05, "loss": 1.7511, "step": 8713 }, { "epoch": 0.9417486220685183, "grad_norm": 0.4200980067253113, "learning_rate": 8e-05, "loss": 1.4749, "step": 8714 }, { "epoch": 0.9418566951259051, "grad_norm": 0.48280343413352966, "learning_rate": 8e-05, "loss": 1.7522, "step": 8715 }, { "epoch": 0.9419647681832919, "grad_norm": 0.46378597617149353, "learning_rate": 8e-05, "loss": 1.6874, "step": 8716 }, { "epoch": 0.9420728412406787, "grad_norm": 0.452905535697937, "learning_rate": 8e-05, "loss": 1.6565, "step": 8717 }, { "epoch": 0.9421809142980655, "grad_norm": 0.4572836458683014, "learning_rate": 8e-05, "loss": 1.6825, "step": 8718 }, { "epoch": 0.9422889873554523, "grad_norm": 0.4696885049343109, "learning_rate": 8e-05, "loss": 1.7248, "step": 8719 }, { "epoch": 0.9423970604128391, "grad_norm": 0.48357516527175903, "learning_rate": 8e-05, "loss": 1.7272, "step": 8720 }, { "epoch": 0.9425051334702259, "grad_norm": 0.4753468930721283, "learning_rate": 8e-05, "loss": 1.6817, "step": 8721 }, { "epoch": 0.9426132065276127, "grad_norm": 0.4739753305912018, "learning_rate": 8e-05, "loss": 1.7342, "step": 8722 }, { "epoch": 0.9427212795849994, "grad_norm": 0.43588685989379883, "learning_rate": 8e-05, "loss": 1.671, "step": 8723 }, { "epoch": 0.9428293526423862, "grad_norm": 0.4845527112483978, "learning_rate": 8e-05, "loss": 1.895, "step": 8724 }, { "epoch": 0.942937425699773, "grad_norm": 0.46178820729255676, "learning_rate": 8e-05, "loss": 1.7151, "step": 8725 }, { "epoch": 0.9430454987571598, "grad_norm": 0.47202056646347046, "learning_rate": 8e-05, "loss": 1.7464, "step": 8726 }, { "epoch": 0.9431535718145466, "grad_norm": 0.4830816984176636, "learning_rate": 8e-05, "loss": 1.7341, "step": 8727 }, { "epoch": 0.9432616448719334, "grad_norm": 0.49544963240623474, "learning_rate": 8e-05, "loss": 1.7428, "step": 8728 }, { "epoch": 0.9433697179293202, "grad_norm": 0.5917497873306274, "learning_rate": 8e-05, "loss": 1.9559, "step": 8729 }, { "epoch": 0.943477790986707, "grad_norm": 0.47523483633995056, "learning_rate": 8e-05, "loss": 1.7815, "step": 8730 }, { "epoch": 0.9435858640440938, "grad_norm": 0.44318222999572754, "learning_rate": 8e-05, "loss": 1.6314, "step": 8731 }, { "epoch": 0.9436939371014806, "grad_norm": 0.4342297911643982, "learning_rate": 8e-05, "loss": 1.485, "step": 8732 }, { "epoch": 0.9438020101588674, "grad_norm": 0.455249547958374, "learning_rate": 8e-05, "loss": 1.6233, "step": 8733 }, { "epoch": 0.9439100832162541, "grad_norm": 0.44273728132247925, "learning_rate": 8e-05, "loss": 1.6493, "step": 8734 }, { "epoch": 0.9440181562736409, "grad_norm": 0.43264368176460266, "learning_rate": 8e-05, "loss": 1.5202, "step": 8735 }, { "epoch": 0.9441262293310277, "grad_norm": 0.439654141664505, "learning_rate": 8e-05, "loss": 1.7138, "step": 8736 }, { "epoch": 0.9442343023884145, "grad_norm": 0.44534045457839966, "learning_rate": 8e-05, "loss": 1.6722, "step": 8737 }, { "epoch": 0.9443423754458014, "grad_norm": 0.4380674660205841, "learning_rate": 8e-05, "loss": 1.5264, "step": 8738 }, { "epoch": 0.9444504485031882, "grad_norm": 0.4387654960155487, "learning_rate": 8e-05, "loss": 1.6509, "step": 8739 }, { "epoch": 0.944558521560575, "grad_norm": 0.4881095588207245, "learning_rate": 8e-05, "loss": 1.8746, "step": 8740 }, { "epoch": 0.9446665946179618, "grad_norm": 0.4441450834274292, "learning_rate": 8e-05, "loss": 1.6264, "step": 8741 }, { "epoch": 0.9447746676753486, "grad_norm": 0.45616769790649414, "learning_rate": 8e-05, "loss": 1.6661, "step": 8742 }, { "epoch": 0.9448827407327354, "grad_norm": 0.4783846437931061, "learning_rate": 8e-05, "loss": 1.7905, "step": 8743 }, { "epoch": 0.9449908137901222, "grad_norm": 0.44875216484069824, "learning_rate": 8e-05, "loss": 1.5695, "step": 8744 }, { "epoch": 0.9450988868475089, "grad_norm": 0.4919162392616272, "learning_rate": 8e-05, "loss": 1.7886, "step": 8745 }, { "epoch": 0.9452069599048957, "grad_norm": 0.4557551443576813, "learning_rate": 8e-05, "loss": 1.7248, "step": 8746 }, { "epoch": 0.9453150329622825, "grad_norm": 0.422952800989151, "learning_rate": 8e-05, "loss": 1.5853, "step": 8747 }, { "epoch": 0.9454231060196693, "grad_norm": 0.422708123922348, "learning_rate": 8e-05, "loss": 1.5253, "step": 8748 }, { "epoch": 0.9455311790770561, "grad_norm": 0.4390639066696167, "learning_rate": 8e-05, "loss": 1.651, "step": 8749 }, { "epoch": 0.9456392521344429, "grad_norm": 0.5084518194198608, "learning_rate": 8e-05, "loss": 1.8834, "step": 8750 }, { "epoch": 0.9457473251918297, "grad_norm": 0.48177066445350647, "learning_rate": 8e-05, "loss": 1.6415, "step": 8751 }, { "epoch": 0.9458553982492165, "grad_norm": 0.4420482814311981, "learning_rate": 8e-05, "loss": 1.6556, "step": 8752 }, { "epoch": 0.9459634713066033, "grad_norm": 0.4703623354434967, "learning_rate": 8e-05, "loss": 1.7267, "step": 8753 }, { "epoch": 0.9460715443639901, "grad_norm": 0.4593774378299713, "learning_rate": 8e-05, "loss": 1.6457, "step": 8754 }, { "epoch": 0.9461796174213769, "grad_norm": 0.46369272470474243, "learning_rate": 8e-05, "loss": 1.6737, "step": 8755 }, { "epoch": 0.9462876904787636, "grad_norm": 0.4876141846179962, "learning_rate": 8e-05, "loss": 1.6589, "step": 8756 }, { "epoch": 0.9463957635361504, "grad_norm": 0.42220914363861084, "learning_rate": 8e-05, "loss": 1.7167, "step": 8757 }, { "epoch": 0.9465038365935372, "grad_norm": 0.5237975716590881, "learning_rate": 8e-05, "loss": 1.59, "step": 8758 }, { "epoch": 0.946611909650924, "grad_norm": 0.4845879077911377, "learning_rate": 8e-05, "loss": 1.6362, "step": 8759 }, { "epoch": 0.9467199827083108, "grad_norm": 0.4851832091808319, "learning_rate": 8e-05, "loss": 1.6506, "step": 8760 }, { "epoch": 0.9468280557656976, "grad_norm": 0.5032719373703003, "learning_rate": 8e-05, "loss": 1.6861, "step": 8761 }, { "epoch": 0.9469361288230844, "grad_norm": 0.45867595076560974, "learning_rate": 8e-05, "loss": 1.7886, "step": 8762 }, { "epoch": 0.9470442018804712, "grad_norm": 0.4751569926738739, "learning_rate": 8e-05, "loss": 1.7258, "step": 8763 }, { "epoch": 0.947152274937858, "grad_norm": 0.46525096893310547, "learning_rate": 8e-05, "loss": 1.6355, "step": 8764 }, { "epoch": 0.9472603479952448, "grad_norm": 0.45295000076293945, "learning_rate": 8e-05, "loss": 1.6967, "step": 8765 }, { "epoch": 0.9473684210526315, "grad_norm": 0.40966305136680603, "learning_rate": 8e-05, "loss": 1.3757, "step": 8766 }, { "epoch": 0.9474764941100183, "grad_norm": 0.4457700252532959, "learning_rate": 8e-05, "loss": 1.6539, "step": 8767 }, { "epoch": 0.9475845671674051, "grad_norm": 0.4716149568557739, "learning_rate": 8e-05, "loss": 1.467, "step": 8768 }, { "epoch": 0.9476926402247919, "grad_norm": 0.4220733642578125, "learning_rate": 8e-05, "loss": 1.5274, "step": 8769 }, { "epoch": 0.9478007132821787, "grad_norm": 0.4724470376968384, "learning_rate": 8e-05, "loss": 1.7335, "step": 8770 }, { "epoch": 0.9479087863395655, "grad_norm": 0.4639819860458374, "learning_rate": 8e-05, "loss": 1.695, "step": 8771 }, { "epoch": 0.9480168593969523, "grad_norm": 0.45211783051490784, "learning_rate": 8e-05, "loss": 1.6997, "step": 8772 }, { "epoch": 0.9481249324543392, "grad_norm": 0.47981637716293335, "learning_rate": 8e-05, "loss": 1.6458, "step": 8773 }, { "epoch": 0.948233005511726, "grad_norm": 0.44600313901901245, "learning_rate": 8e-05, "loss": 1.7135, "step": 8774 }, { "epoch": 0.9483410785691128, "grad_norm": 0.46181604266166687, "learning_rate": 8e-05, "loss": 1.7895, "step": 8775 }, { "epoch": 0.9484491516264996, "grad_norm": 0.48224493861198425, "learning_rate": 8e-05, "loss": 1.7803, "step": 8776 }, { "epoch": 0.9485572246838863, "grad_norm": 0.4706035256385803, "learning_rate": 8e-05, "loss": 1.8015, "step": 8777 }, { "epoch": 0.9486652977412731, "grad_norm": 0.4585378170013428, "learning_rate": 8e-05, "loss": 1.7105, "step": 8778 }, { "epoch": 0.9487733707986599, "grad_norm": 0.4651271104812622, "learning_rate": 8e-05, "loss": 1.7681, "step": 8779 }, { "epoch": 0.9488814438560467, "grad_norm": 0.48164045810699463, "learning_rate": 8e-05, "loss": 1.6912, "step": 8780 }, { "epoch": 0.9489895169134335, "grad_norm": 0.4242510497570038, "learning_rate": 8e-05, "loss": 1.5014, "step": 8781 }, { "epoch": 0.9490975899708203, "grad_norm": 0.43801379203796387, "learning_rate": 8e-05, "loss": 1.6121, "step": 8782 }, { "epoch": 0.9492056630282071, "grad_norm": 0.45008331537246704, "learning_rate": 8e-05, "loss": 1.5937, "step": 8783 }, { "epoch": 0.9493137360855939, "grad_norm": 0.44011542201042175, "learning_rate": 8e-05, "loss": 1.6616, "step": 8784 }, { "epoch": 0.9494218091429807, "grad_norm": 0.4892531633377075, "learning_rate": 8e-05, "loss": 1.8194, "step": 8785 }, { "epoch": 0.9495298822003675, "grad_norm": 0.4646034240722656, "learning_rate": 8e-05, "loss": 1.6328, "step": 8786 }, { "epoch": 0.9496379552577543, "grad_norm": 0.45697954297065735, "learning_rate": 8e-05, "loss": 1.6377, "step": 8787 }, { "epoch": 0.949746028315141, "grad_norm": 0.4328789710998535, "learning_rate": 8e-05, "loss": 1.5207, "step": 8788 }, { "epoch": 0.9498541013725278, "grad_norm": 0.4558294713497162, "learning_rate": 8e-05, "loss": 1.6449, "step": 8789 }, { "epoch": 0.9499621744299146, "grad_norm": 0.44615495204925537, "learning_rate": 8e-05, "loss": 1.7849, "step": 8790 }, { "epoch": 0.9500702474873014, "grad_norm": 0.500220000743866, "learning_rate": 8e-05, "loss": 1.661, "step": 8791 }, { "epoch": 0.9501783205446882, "grad_norm": 0.43266022205352783, "learning_rate": 8e-05, "loss": 1.6408, "step": 8792 }, { "epoch": 0.950286393602075, "grad_norm": 0.4563833177089691, "learning_rate": 8e-05, "loss": 1.622, "step": 8793 }, { "epoch": 0.9503944666594618, "grad_norm": 0.4770665764808655, "learning_rate": 8e-05, "loss": 1.7774, "step": 8794 }, { "epoch": 0.9505025397168486, "grad_norm": 0.47093498706817627, "learning_rate": 8e-05, "loss": 1.7618, "step": 8795 }, { "epoch": 0.9506106127742354, "grad_norm": 0.4519962966442108, "learning_rate": 8e-05, "loss": 1.6257, "step": 8796 }, { "epoch": 0.9507186858316222, "grad_norm": 0.41748666763305664, "learning_rate": 8e-05, "loss": 1.6657, "step": 8797 }, { "epoch": 0.9508267588890089, "grad_norm": 0.46328267455101013, "learning_rate": 8e-05, "loss": 1.7265, "step": 8798 }, { "epoch": 0.9509348319463957, "grad_norm": 0.43140843510627747, "learning_rate": 8e-05, "loss": 1.6804, "step": 8799 }, { "epoch": 0.9510429050037825, "grad_norm": 0.4223465621471405, "learning_rate": 8e-05, "loss": 1.5577, "step": 8800 }, { "epoch": 0.9511509780611693, "grad_norm": 0.4555056393146515, "learning_rate": 8e-05, "loss": 1.6436, "step": 8801 }, { "epoch": 0.9512590511185561, "grad_norm": 0.4495524764060974, "learning_rate": 8e-05, "loss": 1.7978, "step": 8802 }, { "epoch": 0.9513671241759429, "grad_norm": 0.46092939376831055, "learning_rate": 8e-05, "loss": 1.7258, "step": 8803 }, { "epoch": 0.9514751972333297, "grad_norm": 0.4680089056491852, "learning_rate": 8e-05, "loss": 1.685, "step": 8804 }, { "epoch": 0.9515832702907165, "grad_norm": 0.43791258335113525, "learning_rate": 8e-05, "loss": 1.5546, "step": 8805 }, { "epoch": 0.9516913433481033, "grad_norm": 0.44161736965179443, "learning_rate": 8e-05, "loss": 1.6461, "step": 8806 }, { "epoch": 0.9517994164054902, "grad_norm": 0.47673219442367554, "learning_rate": 8e-05, "loss": 1.7719, "step": 8807 }, { "epoch": 0.951907489462877, "grad_norm": 0.4702881872653961, "learning_rate": 8e-05, "loss": 1.7516, "step": 8808 }, { "epoch": 0.9520155625202636, "grad_norm": 0.4692835211753845, "learning_rate": 8e-05, "loss": 1.7789, "step": 8809 }, { "epoch": 0.9521236355776505, "grad_norm": 0.44515490531921387, "learning_rate": 8e-05, "loss": 1.7403, "step": 8810 }, { "epoch": 0.9522317086350373, "grad_norm": 0.4720122814178467, "learning_rate": 8e-05, "loss": 1.7607, "step": 8811 }, { "epoch": 0.9523397816924241, "grad_norm": 0.4280131459236145, "learning_rate": 8e-05, "loss": 1.6511, "step": 8812 }, { "epoch": 0.9524478547498109, "grad_norm": 0.43153059482574463, "learning_rate": 8e-05, "loss": 1.638, "step": 8813 }, { "epoch": 0.9525559278071977, "grad_norm": 0.44257357716560364, "learning_rate": 8e-05, "loss": 1.709, "step": 8814 }, { "epoch": 0.9526640008645845, "grad_norm": 0.4583471119403839, "learning_rate": 8e-05, "loss": 1.5832, "step": 8815 }, { "epoch": 0.9527720739219713, "grad_norm": 0.4964755177497864, "learning_rate": 8e-05, "loss": 1.6133, "step": 8816 }, { "epoch": 0.9528801469793581, "grad_norm": 0.5241087079048157, "learning_rate": 8e-05, "loss": 1.8749, "step": 8817 }, { "epoch": 0.9529882200367449, "grad_norm": 0.48036089539527893, "learning_rate": 8e-05, "loss": 1.6893, "step": 8818 }, { "epoch": 0.9530962930941317, "grad_norm": 0.43999406695365906, "learning_rate": 8e-05, "loss": 1.7531, "step": 8819 }, { "epoch": 0.9532043661515184, "grad_norm": 0.5101463794708252, "learning_rate": 8e-05, "loss": 1.9741, "step": 8820 }, { "epoch": 0.9533124392089052, "grad_norm": 0.49088194966316223, "learning_rate": 8e-05, "loss": 1.7227, "step": 8821 }, { "epoch": 0.953420512266292, "grad_norm": 0.5063536763191223, "learning_rate": 8e-05, "loss": 1.9238, "step": 8822 }, { "epoch": 0.9535285853236788, "grad_norm": 0.4529161751270294, "learning_rate": 8e-05, "loss": 1.7331, "step": 8823 }, { "epoch": 0.9536366583810656, "grad_norm": 0.47761833667755127, "learning_rate": 8e-05, "loss": 1.7843, "step": 8824 }, { "epoch": 0.9537447314384524, "grad_norm": 0.4524553120136261, "learning_rate": 8e-05, "loss": 1.6117, "step": 8825 }, { "epoch": 0.9538528044958392, "grad_norm": 0.42959603667259216, "learning_rate": 8e-05, "loss": 1.6281, "step": 8826 }, { "epoch": 0.953960877553226, "grad_norm": 0.43870916962623596, "learning_rate": 8e-05, "loss": 1.649, "step": 8827 }, { "epoch": 0.9540689506106128, "grad_norm": 0.4297657310962677, "learning_rate": 8e-05, "loss": 1.6236, "step": 8828 }, { "epoch": 0.9541770236679996, "grad_norm": 0.46857285499572754, "learning_rate": 8e-05, "loss": 1.7974, "step": 8829 }, { "epoch": 0.9542850967253863, "grad_norm": 0.5468045473098755, "learning_rate": 8e-05, "loss": 1.9444, "step": 8830 }, { "epoch": 0.9543931697827731, "grad_norm": 0.4775119721889496, "learning_rate": 8e-05, "loss": 1.7397, "step": 8831 }, { "epoch": 0.9545012428401599, "grad_norm": 0.4600176215171814, "learning_rate": 8e-05, "loss": 1.6103, "step": 8832 }, { "epoch": 0.9546093158975467, "grad_norm": 0.4495052397251129, "learning_rate": 8e-05, "loss": 1.7323, "step": 8833 }, { "epoch": 0.9547173889549335, "grad_norm": 0.4204075038433075, "learning_rate": 8e-05, "loss": 1.5219, "step": 8834 }, { "epoch": 0.9548254620123203, "grad_norm": 0.4611603617668152, "learning_rate": 8e-05, "loss": 1.6886, "step": 8835 }, { "epoch": 0.9549335350697071, "grad_norm": 0.41263630986213684, "learning_rate": 8e-05, "loss": 1.6587, "step": 8836 }, { "epoch": 0.9550416081270939, "grad_norm": 0.46740400791168213, "learning_rate": 8e-05, "loss": 1.6513, "step": 8837 }, { "epoch": 0.9551496811844807, "grad_norm": 0.4199559688568115, "learning_rate": 8e-05, "loss": 1.6052, "step": 8838 }, { "epoch": 0.9552577542418675, "grad_norm": 0.4528767168521881, "learning_rate": 8e-05, "loss": 1.7387, "step": 8839 }, { "epoch": 0.9553658272992543, "grad_norm": 0.5142785906791687, "learning_rate": 8e-05, "loss": 2.0406, "step": 8840 }, { "epoch": 0.955473900356641, "grad_norm": 0.4415123462677002, "learning_rate": 8e-05, "loss": 1.5267, "step": 8841 }, { "epoch": 0.9555819734140278, "grad_norm": 0.47243908047676086, "learning_rate": 8e-05, "loss": 1.6808, "step": 8842 }, { "epoch": 0.9556900464714146, "grad_norm": 0.5008917450904846, "learning_rate": 8e-05, "loss": 1.5794, "step": 8843 }, { "epoch": 0.9557981195288014, "grad_norm": 0.4796378016471863, "learning_rate": 8e-05, "loss": 1.6958, "step": 8844 }, { "epoch": 0.9559061925861883, "grad_norm": 0.4493555724620819, "learning_rate": 8e-05, "loss": 1.4794, "step": 8845 }, { "epoch": 0.9560142656435751, "grad_norm": 0.40745270252227783, "learning_rate": 8e-05, "loss": 1.5036, "step": 8846 }, { "epoch": 0.9561223387009619, "grad_norm": 0.4769785702228546, "learning_rate": 8e-05, "loss": 1.8166, "step": 8847 }, { "epoch": 0.9562304117583487, "grad_norm": 0.4279644787311554, "learning_rate": 8e-05, "loss": 1.6541, "step": 8848 }, { "epoch": 0.9563384848157355, "grad_norm": 0.5131330490112305, "learning_rate": 8e-05, "loss": 1.7701, "step": 8849 }, { "epoch": 0.9564465578731223, "grad_norm": 0.4360217750072479, "learning_rate": 8e-05, "loss": 1.5872, "step": 8850 }, { "epoch": 0.9565546309305091, "grad_norm": 0.4384888708591461, "learning_rate": 8e-05, "loss": 1.5955, "step": 8851 }, { "epoch": 0.9566627039878958, "grad_norm": 0.5086166262626648, "learning_rate": 8e-05, "loss": 1.8857, "step": 8852 }, { "epoch": 0.9567707770452826, "grad_norm": 0.42783647775650024, "learning_rate": 8e-05, "loss": 1.6343, "step": 8853 }, { "epoch": 0.9568788501026694, "grad_norm": 0.47867268323898315, "learning_rate": 8e-05, "loss": 1.7086, "step": 8854 }, { "epoch": 0.9569869231600562, "grad_norm": 0.480796754360199, "learning_rate": 8e-05, "loss": 1.8674, "step": 8855 }, { "epoch": 0.957094996217443, "grad_norm": 0.46404922008514404, "learning_rate": 8e-05, "loss": 1.5544, "step": 8856 }, { "epoch": 0.9572030692748298, "grad_norm": 0.4589596688747406, "learning_rate": 8e-05, "loss": 1.779, "step": 8857 }, { "epoch": 0.9573111423322166, "grad_norm": 0.4727623760700226, "learning_rate": 8e-05, "loss": 1.6652, "step": 8858 }, { "epoch": 0.9574192153896034, "grad_norm": 0.4673205316066742, "learning_rate": 8e-05, "loss": 1.7486, "step": 8859 }, { "epoch": 0.9575272884469902, "grad_norm": 0.4839211106300354, "learning_rate": 8e-05, "loss": 1.6949, "step": 8860 }, { "epoch": 0.957635361504377, "grad_norm": 0.46019822359085083, "learning_rate": 8e-05, "loss": 1.5588, "step": 8861 }, { "epoch": 0.9577434345617638, "grad_norm": 0.448860764503479, "learning_rate": 8e-05, "loss": 1.6515, "step": 8862 }, { "epoch": 0.9578515076191505, "grad_norm": 0.4316478371620178, "learning_rate": 8e-05, "loss": 1.5952, "step": 8863 }, { "epoch": 0.9579595806765373, "grad_norm": 0.4384631812572479, "learning_rate": 8e-05, "loss": 1.7154, "step": 8864 }, { "epoch": 0.9580676537339241, "grad_norm": 0.4798904061317444, "learning_rate": 8e-05, "loss": 1.7086, "step": 8865 }, { "epoch": 0.9581757267913109, "grad_norm": 0.42315930128097534, "learning_rate": 8e-05, "loss": 1.5535, "step": 8866 }, { "epoch": 0.9582837998486977, "grad_norm": 0.4418141543865204, "learning_rate": 8e-05, "loss": 1.6185, "step": 8867 }, { "epoch": 0.9583918729060845, "grad_norm": 0.41637521982192993, "learning_rate": 8e-05, "loss": 1.591, "step": 8868 }, { "epoch": 0.9584999459634713, "grad_norm": 0.43025097250938416, "learning_rate": 8e-05, "loss": 1.5664, "step": 8869 }, { "epoch": 0.9586080190208581, "grad_norm": 0.439468652009964, "learning_rate": 8e-05, "loss": 1.6488, "step": 8870 }, { "epoch": 0.9587160920782449, "grad_norm": 0.4342207908630371, "learning_rate": 8e-05, "loss": 1.6895, "step": 8871 }, { "epoch": 0.9588241651356317, "grad_norm": 0.43364715576171875, "learning_rate": 8e-05, "loss": 1.6794, "step": 8872 }, { "epoch": 0.9589322381930184, "grad_norm": 0.49756374955177307, "learning_rate": 8e-05, "loss": 1.8369, "step": 8873 }, { "epoch": 0.9590403112504052, "grad_norm": 0.45102307200431824, "learning_rate": 8e-05, "loss": 1.4895, "step": 8874 }, { "epoch": 0.959148384307792, "grad_norm": 0.44980114698410034, "learning_rate": 8e-05, "loss": 1.6767, "step": 8875 }, { "epoch": 0.9592564573651788, "grad_norm": 0.45605549216270447, "learning_rate": 8e-05, "loss": 1.7248, "step": 8876 }, { "epoch": 0.9593645304225656, "grad_norm": 0.47105348110198975, "learning_rate": 8e-05, "loss": 1.7565, "step": 8877 }, { "epoch": 0.9594726034799524, "grad_norm": 0.43229615688323975, "learning_rate": 8e-05, "loss": 1.6586, "step": 8878 }, { "epoch": 0.9595806765373392, "grad_norm": 0.43487808108329773, "learning_rate": 8e-05, "loss": 1.6176, "step": 8879 }, { "epoch": 0.959688749594726, "grad_norm": 0.48350492119789124, "learning_rate": 8e-05, "loss": 1.7675, "step": 8880 }, { "epoch": 0.9597968226521129, "grad_norm": 0.4513801038265228, "learning_rate": 8e-05, "loss": 1.6305, "step": 8881 }, { "epoch": 0.9599048957094997, "grad_norm": 0.4690123200416565, "learning_rate": 8e-05, "loss": 1.6633, "step": 8882 }, { "epoch": 0.9600129687668865, "grad_norm": 0.45219042897224426, "learning_rate": 8e-05, "loss": 1.5727, "step": 8883 }, { "epoch": 0.9601210418242732, "grad_norm": 0.4427321255207062, "learning_rate": 8e-05, "loss": 1.6356, "step": 8884 }, { "epoch": 0.96022911488166, "grad_norm": 0.44588223099708557, "learning_rate": 8e-05, "loss": 1.6205, "step": 8885 }, { "epoch": 0.9603371879390468, "grad_norm": 0.42701438069343567, "learning_rate": 8e-05, "loss": 1.647, "step": 8886 }, { "epoch": 0.9604452609964336, "grad_norm": 0.43103542923927307, "learning_rate": 8e-05, "loss": 1.5278, "step": 8887 }, { "epoch": 0.9605533340538204, "grad_norm": 0.4841434359550476, "learning_rate": 8e-05, "loss": 1.8926, "step": 8888 }, { "epoch": 0.9606614071112072, "grad_norm": 0.4974484443664551, "learning_rate": 8e-05, "loss": 1.6118, "step": 8889 }, { "epoch": 0.960769480168594, "grad_norm": 0.44634953141212463, "learning_rate": 8e-05, "loss": 1.4439, "step": 8890 }, { "epoch": 0.9608775532259808, "grad_norm": 0.46460670232772827, "learning_rate": 8e-05, "loss": 1.4108, "step": 8891 }, { "epoch": 0.9609856262833676, "grad_norm": 0.4747583568096161, "learning_rate": 8e-05, "loss": 1.783, "step": 8892 }, { "epoch": 0.9610936993407544, "grad_norm": 0.4782462418079376, "learning_rate": 8e-05, "loss": 1.6477, "step": 8893 }, { "epoch": 0.9612017723981412, "grad_norm": 0.44737309217453003, "learning_rate": 8e-05, "loss": 1.7468, "step": 8894 }, { "epoch": 0.9613098454555279, "grad_norm": 0.4353908598423004, "learning_rate": 8e-05, "loss": 1.7475, "step": 8895 }, { "epoch": 0.9614179185129147, "grad_norm": 0.42825639247894287, "learning_rate": 8e-05, "loss": 1.6247, "step": 8896 }, { "epoch": 0.9615259915703015, "grad_norm": 0.48314833641052246, "learning_rate": 8e-05, "loss": 1.7953, "step": 8897 }, { "epoch": 0.9616340646276883, "grad_norm": 0.4190172255039215, "learning_rate": 8e-05, "loss": 1.6067, "step": 8898 }, { "epoch": 0.9617421376850751, "grad_norm": 0.42906561493873596, "learning_rate": 8e-05, "loss": 1.5405, "step": 8899 }, { "epoch": 0.9618502107424619, "grad_norm": 0.45301809906959534, "learning_rate": 8e-05, "loss": 1.6092, "step": 8900 }, { "epoch": 0.9619582837998487, "grad_norm": 0.46587073802948, "learning_rate": 8e-05, "loss": 1.7808, "step": 8901 }, { "epoch": 0.9620663568572355, "grad_norm": 0.42377302050590515, "learning_rate": 8e-05, "loss": 1.4956, "step": 8902 }, { "epoch": 0.9621744299146223, "grad_norm": 0.45605766773223877, "learning_rate": 8e-05, "loss": 1.7098, "step": 8903 }, { "epoch": 0.9622825029720091, "grad_norm": 0.47951844334602356, "learning_rate": 8e-05, "loss": 1.6054, "step": 8904 }, { "epoch": 0.9623905760293958, "grad_norm": 0.4786067008972168, "learning_rate": 8e-05, "loss": 1.7158, "step": 8905 }, { "epoch": 0.9624986490867826, "grad_norm": 0.4312930703163147, "learning_rate": 8e-05, "loss": 1.6247, "step": 8906 }, { "epoch": 0.9626067221441694, "grad_norm": 0.4520036280155182, "learning_rate": 8e-05, "loss": 1.7454, "step": 8907 }, { "epoch": 0.9627147952015562, "grad_norm": 0.4750392735004425, "learning_rate": 8e-05, "loss": 1.6184, "step": 8908 }, { "epoch": 0.962822868258943, "grad_norm": 0.48487985134124756, "learning_rate": 8e-05, "loss": 1.6497, "step": 8909 }, { "epoch": 0.9629309413163298, "grad_norm": 0.5306264758110046, "learning_rate": 8e-05, "loss": 1.8252, "step": 8910 }, { "epoch": 0.9630390143737166, "grad_norm": 0.45690372586250305, "learning_rate": 8e-05, "loss": 1.801, "step": 8911 }, { "epoch": 0.9631470874311034, "grad_norm": 0.45286184549331665, "learning_rate": 8e-05, "loss": 1.6561, "step": 8912 }, { "epoch": 0.9632551604884902, "grad_norm": 0.46927353739738464, "learning_rate": 8e-05, "loss": 1.7902, "step": 8913 }, { "epoch": 0.963363233545877, "grad_norm": 0.4218420386314392, "learning_rate": 8e-05, "loss": 1.6493, "step": 8914 }, { "epoch": 0.9634713066032639, "grad_norm": 0.44368574023246765, "learning_rate": 8e-05, "loss": 1.4393, "step": 8915 }, { "epoch": 0.9635793796606505, "grad_norm": 0.42260706424713135, "learning_rate": 8e-05, "loss": 1.4508, "step": 8916 }, { "epoch": 0.9636874527180374, "grad_norm": 0.5253618955612183, "learning_rate": 8e-05, "loss": 1.8045, "step": 8917 }, { "epoch": 0.9637955257754242, "grad_norm": 0.43795397877693176, "learning_rate": 8e-05, "loss": 1.7437, "step": 8918 }, { "epoch": 0.963903598832811, "grad_norm": 0.47467315196990967, "learning_rate": 8e-05, "loss": 1.7083, "step": 8919 }, { "epoch": 0.9640116718901978, "grad_norm": 0.4554799795150757, "learning_rate": 8e-05, "loss": 1.6751, "step": 8920 }, { "epoch": 0.9641197449475846, "grad_norm": 0.4579152762889862, "learning_rate": 8e-05, "loss": 1.6238, "step": 8921 }, { "epoch": 0.9642278180049714, "grad_norm": 0.42992958426475525, "learning_rate": 8e-05, "loss": 1.6695, "step": 8922 }, { "epoch": 0.9643358910623582, "grad_norm": 0.4474663734436035, "learning_rate": 8e-05, "loss": 1.6548, "step": 8923 }, { "epoch": 0.964443964119745, "grad_norm": 0.4591861069202423, "learning_rate": 8e-05, "loss": 1.6632, "step": 8924 }, { "epoch": 0.9645520371771318, "grad_norm": 0.46306875348091125, "learning_rate": 8e-05, "loss": 1.878, "step": 8925 }, { "epoch": 0.9646601102345186, "grad_norm": 0.4421653151512146, "learning_rate": 8e-05, "loss": 1.5801, "step": 8926 }, { "epoch": 0.9647681832919053, "grad_norm": 0.4390055537223816, "learning_rate": 8e-05, "loss": 1.6226, "step": 8927 }, { "epoch": 0.9648762563492921, "grad_norm": 0.4160687327384949, "learning_rate": 8e-05, "loss": 1.6414, "step": 8928 }, { "epoch": 0.9649843294066789, "grad_norm": 0.4217209815979004, "learning_rate": 8e-05, "loss": 1.5468, "step": 8929 }, { "epoch": 0.9650924024640657, "grad_norm": 0.4664812982082367, "learning_rate": 8e-05, "loss": 1.7961, "step": 8930 }, { "epoch": 0.9652004755214525, "grad_norm": 0.45663127303123474, "learning_rate": 8e-05, "loss": 1.7379, "step": 8931 }, { "epoch": 0.9653085485788393, "grad_norm": 0.45595118403434753, "learning_rate": 8e-05, "loss": 1.8044, "step": 8932 }, { "epoch": 0.9654166216362261, "grad_norm": 0.4720596969127655, "learning_rate": 8e-05, "loss": 1.8752, "step": 8933 }, { "epoch": 0.9655246946936129, "grad_norm": 0.4044124186038971, "learning_rate": 8e-05, "loss": 1.5391, "step": 8934 }, { "epoch": 0.9656327677509997, "grad_norm": 0.4237770438194275, "learning_rate": 8e-05, "loss": 1.5786, "step": 8935 }, { "epoch": 0.9657408408083865, "grad_norm": 0.47881385684013367, "learning_rate": 8e-05, "loss": 1.5113, "step": 8936 }, { "epoch": 0.9658489138657733, "grad_norm": 0.4938980042934418, "learning_rate": 8e-05, "loss": 1.7608, "step": 8937 }, { "epoch": 0.96595698692316, "grad_norm": 0.44691309332847595, "learning_rate": 8e-05, "loss": 1.6187, "step": 8938 }, { "epoch": 0.9660650599805468, "grad_norm": 0.4321226179599762, "learning_rate": 8e-05, "loss": 1.6471, "step": 8939 }, { "epoch": 0.9661731330379336, "grad_norm": 0.4889061748981476, "learning_rate": 8e-05, "loss": 1.7462, "step": 8940 }, { "epoch": 0.9662812060953204, "grad_norm": 0.47289079427719116, "learning_rate": 8e-05, "loss": 1.716, "step": 8941 }, { "epoch": 0.9663892791527072, "grad_norm": 0.4880748987197876, "learning_rate": 8e-05, "loss": 1.7584, "step": 8942 }, { "epoch": 0.966497352210094, "grad_norm": 0.4177621304988861, "learning_rate": 8e-05, "loss": 1.566, "step": 8943 }, { "epoch": 0.9666054252674808, "grad_norm": 0.4447115659713745, "learning_rate": 8e-05, "loss": 1.7153, "step": 8944 }, { "epoch": 0.9667134983248676, "grad_norm": 0.41890841722488403, "learning_rate": 8e-05, "loss": 1.6278, "step": 8945 }, { "epoch": 0.9668215713822544, "grad_norm": 0.4454350173473358, "learning_rate": 8e-05, "loss": 1.6821, "step": 8946 }, { "epoch": 0.9669296444396412, "grad_norm": 0.4871027171611786, "learning_rate": 8e-05, "loss": 1.6178, "step": 8947 }, { "epoch": 0.9670377174970279, "grad_norm": 0.42066383361816406, "learning_rate": 8e-05, "loss": 1.7406, "step": 8948 }, { "epoch": 0.9671457905544147, "grad_norm": 0.4661374092102051, "learning_rate": 8e-05, "loss": 1.6734, "step": 8949 }, { "epoch": 0.9672538636118015, "grad_norm": 0.46928325295448303, "learning_rate": 8e-05, "loss": 1.8682, "step": 8950 }, { "epoch": 0.9673619366691883, "grad_norm": 0.520987331867218, "learning_rate": 8e-05, "loss": 1.609, "step": 8951 }, { "epoch": 0.9674700097265752, "grad_norm": 0.4803849756717682, "learning_rate": 8e-05, "loss": 1.7557, "step": 8952 }, { "epoch": 0.967578082783962, "grad_norm": 0.45628365874290466, "learning_rate": 8e-05, "loss": 1.4611, "step": 8953 }, { "epoch": 0.9676861558413488, "grad_norm": 0.4447997510433197, "learning_rate": 8e-05, "loss": 1.6887, "step": 8954 }, { "epoch": 0.9677942288987356, "grad_norm": 0.4339659810066223, "learning_rate": 8e-05, "loss": 1.6256, "step": 8955 }, { "epoch": 0.9679023019561224, "grad_norm": 0.4820590019226074, "learning_rate": 8e-05, "loss": 1.7091, "step": 8956 }, { "epoch": 0.9680103750135092, "grad_norm": 0.45699596405029297, "learning_rate": 8e-05, "loss": 1.5442, "step": 8957 }, { "epoch": 0.968118448070896, "grad_norm": 0.46150150895118713, "learning_rate": 8e-05, "loss": 1.6124, "step": 8958 }, { "epoch": 0.9682265211282827, "grad_norm": 0.4300921857357025, "learning_rate": 8e-05, "loss": 1.7055, "step": 8959 }, { "epoch": 0.9683345941856695, "grad_norm": 0.43874219059944153, "learning_rate": 8e-05, "loss": 1.5469, "step": 8960 }, { "epoch": 0.9684426672430563, "grad_norm": 0.47345876693725586, "learning_rate": 8e-05, "loss": 1.8607, "step": 8961 }, { "epoch": 0.9685507403004431, "grad_norm": 0.45129865407943726, "learning_rate": 8e-05, "loss": 1.5578, "step": 8962 }, { "epoch": 0.9686588133578299, "grad_norm": 0.48045802116394043, "learning_rate": 8e-05, "loss": 1.6563, "step": 8963 }, { "epoch": 0.9687668864152167, "grad_norm": 0.42299550771713257, "learning_rate": 8e-05, "loss": 1.5809, "step": 8964 }, { "epoch": 0.9688749594726035, "grad_norm": 0.45808273553848267, "learning_rate": 8e-05, "loss": 1.7833, "step": 8965 }, { "epoch": 0.9689830325299903, "grad_norm": 0.42783284187316895, "learning_rate": 8e-05, "loss": 1.6036, "step": 8966 }, { "epoch": 0.9690911055873771, "grad_norm": 0.4389735758304596, "learning_rate": 8e-05, "loss": 1.7913, "step": 8967 }, { "epoch": 0.9691991786447639, "grad_norm": 0.46372750401496887, "learning_rate": 8e-05, "loss": 1.6063, "step": 8968 }, { "epoch": 0.9693072517021507, "grad_norm": 0.4240957498550415, "learning_rate": 8e-05, "loss": 1.5132, "step": 8969 }, { "epoch": 0.9694153247595374, "grad_norm": 0.5247745513916016, "learning_rate": 8e-05, "loss": 1.7563, "step": 8970 }, { "epoch": 0.9695233978169242, "grad_norm": 0.4999253451824188, "learning_rate": 8e-05, "loss": 1.9259, "step": 8971 }, { "epoch": 0.969631470874311, "grad_norm": 0.44211316108703613, "learning_rate": 8e-05, "loss": 1.5707, "step": 8972 }, { "epoch": 0.9697395439316978, "grad_norm": 0.45343658328056335, "learning_rate": 8e-05, "loss": 1.6428, "step": 8973 }, { "epoch": 0.9698476169890846, "grad_norm": 0.4524272680282593, "learning_rate": 8e-05, "loss": 1.6459, "step": 8974 }, { "epoch": 0.9699556900464714, "grad_norm": 0.449339896440506, "learning_rate": 8e-05, "loss": 1.6714, "step": 8975 }, { "epoch": 0.9700637631038582, "grad_norm": 0.4514801800251007, "learning_rate": 8e-05, "loss": 1.6256, "step": 8976 }, { "epoch": 0.970171836161245, "grad_norm": 0.4480835497379303, "learning_rate": 8e-05, "loss": 1.7419, "step": 8977 }, { "epoch": 0.9702799092186318, "grad_norm": 0.45284661650657654, "learning_rate": 8e-05, "loss": 1.7297, "step": 8978 }, { "epoch": 0.9703879822760186, "grad_norm": 0.4401254653930664, "learning_rate": 8e-05, "loss": 1.7158, "step": 8979 }, { "epoch": 0.9704960553334053, "grad_norm": 0.4497932195663452, "learning_rate": 8e-05, "loss": 1.6216, "step": 8980 }, { "epoch": 0.9706041283907921, "grad_norm": 0.4820152521133423, "learning_rate": 8e-05, "loss": 1.7116, "step": 8981 }, { "epoch": 0.9707122014481789, "grad_norm": 0.43661317229270935, "learning_rate": 8e-05, "loss": 1.6951, "step": 8982 }, { "epoch": 0.9708202745055657, "grad_norm": 0.4642453193664551, "learning_rate": 8e-05, "loss": 1.7768, "step": 8983 }, { "epoch": 0.9709283475629525, "grad_norm": 0.4369053840637207, "learning_rate": 8e-05, "loss": 1.6207, "step": 8984 }, { "epoch": 0.9710364206203393, "grad_norm": 0.42287296056747437, "learning_rate": 8e-05, "loss": 1.5815, "step": 8985 }, { "epoch": 0.9711444936777262, "grad_norm": 0.4180700480937958, "learning_rate": 8e-05, "loss": 1.596, "step": 8986 }, { "epoch": 0.971252566735113, "grad_norm": 0.46138960123062134, "learning_rate": 8e-05, "loss": 1.6632, "step": 8987 }, { "epoch": 0.9713606397924998, "grad_norm": 0.4232023060321808, "learning_rate": 8e-05, "loss": 1.6676, "step": 8988 }, { "epoch": 0.9714687128498866, "grad_norm": 0.472800076007843, "learning_rate": 8e-05, "loss": 1.8307, "step": 8989 }, { "epoch": 0.9715767859072734, "grad_norm": 0.4579298794269562, "learning_rate": 8e-05, "loss": 1.7552, "step": 8990 }, { "epoch": 0.9716848589646601, "grad_norm": 0.4849993586540222, "learning_rate": 8e-05, "loss": 1.5284, "step": 8991 }, { "epoch": 0.9717929320220469, "grad_norm": 0.4971640408039093, "learning_rate": 8e-05, "loss": 1.6878, "step": 8992 }, { "epoch": 0.9719010050794337, "grad_norm": 0.439851850271225, "learning_rate": 8e-05, "loss": 1.5847, "step": 8993 }, { "epoch": 0.9720090781368205, "grad_norm": 0.47452253103256226, "learning_rate": 8e-05, "loss": 1.6695, "step": 8994 }, { "epoch": 0.9721171511942073, "grad_norm": 0.4631219804286957, "learning_rate": 8e-05, "loss": 1.7367, "step": 8995 }, { "epoch": 0.9722252242515941, "grad_norm": 0.45604726672172546, "learning_rate": 8e-05, "loss": 1.7526, "step": 8996 }, { "epoch": 0.9723332973089809, "grad_norm": 0.4581926465034485, "learning_rate": 8e-05, "loss": 1.5905, "step": 8997 }, { "epoch": 0.9724413703663677, "grad_norm": 0.462313175201416, "learning_rate": 8e-05, "loss": 1.5074, "step": 8998 }, { "epoch": 0.9725494434237545, "grad_norm": 0.4445895552635193, "learning_rate": 8e-05, "loss": 1.6735, "step": 8999 }, { "epoch": 0.9726575164811413, "grad_norm": 0.4367079436779022, "learning_rate": 8e-05, "loss": 1.6524, "step": 9000 }, { "epoch": 0.9727655895385281, "grad_norm": 0.4256881773471832, "learning_rate": 8e-05, "loss": 1.4674, "step": 9001 }, { "epoch": 0.9728736625959148, "grad_norm": 0.46901366114616394, "learning_rate": 8e-05, "loss": 1.5329, "step": 9002 }, { "epoch": 0.9729817356533016, "grad_norm": 0.4626148045063019, "learning_rate": 8e-05, "loss": 1.684, "step": 9003 }, { "epoch": 0.9730898087106884, "grad_norm": 0.44191303849220276, "learning_rate": 8e-05, "loss": 1.679, "step": 9004 }, { "epoch": 0.9731978817680752, "grad_norm": 0.49317046999931335, "learning_rate": 8e-05, "loss": 1.6711, "step": 9005 }, { "epoch": 0.973305954825462, "grad_norm": 0.45146504044532776, "learning_rate": 8e-05, "loss": 1.636, "step": 9006 }, { "epoch": 0.9734140278828488, "grad_norm": 0.431815505027771, "learning_rate": 8e-05, "loss": 1.7929, "step": 9007 }, { "epoch": 0.9735221009402356, "grad_norm": 0.46657437086105347, "learning_rate": 8e-05, "loss": 1.5364, "step": 9008 }, { "epoch": 0.9736301739976224, "grad_norm": 0.45548319816589355, "learning_rate": 8e-05, "loss": 1.7308, "step": 9009 }, { "epoch": 0.9737382470550092, "grad_norm": 0.46613261103630066, "learning_rate": 8e-05, "loss": 1.6224, "step": 9010 }, { "epoch": 0.973846320112396, "grad_norm": 0.4994349181652069, "learning_rate": 8e-05, "loss": 1.6896, "step": 9011 }, { "epoch": 0.9739543931697828, "grad_norm": 0.4544588327407837, "learning_rate": 8e-05, "loss": 1.535, "step": 9012 }, { "epoch": 0.9740624662271695, "grad_norm": 0.5024668574333191, "learning_rate": 8e-05, "loss": 1.883, "step": 9013 }, { "epoch": 0.9741705392845563, "grad_norm": 0.43656212091445923, "learning_rate": 8e-05, "loss": 1.6136, "step": 9014 }, { "epoch": 0.9742786123419431, "grad_norm": 0.49243655800819397, "learning_rate": 8e-05, "loss": 1.8807, "step": 9015 }, { "epoch": 0.9743866853993299, "grad_norm": 0.4667496979236603, "learning_rate": 8e-05, "loss": 1.7536, "step": 9016 }, { "epoch": 0.9744947584567167, "grad_norm": 0.5002925992012024, "learning_rate": 8e-05, "loss": 1.425, "step": 9017 }, { "epoch": 0.9746028315141035, "grad_norm": 0.5787729620933533, "learning_rate": 8e-05, "loss": 2.0132, "step": 9018 }, { "epoch": 0.9747109045714903, "grad_norm": 0.46440058946609497, "learning_rate": 8e-05, "loss": 1.6552, "step": 9019 }, { "epoch": 0.9748189776288771, "grad_norm": 0.40610629320144653, "learning_rate": 8e-05, "loss": 1.6277, "step": 9020 }, { "epoch": 0.974927050686264, "grad_norm": 0.44509464502334595, "learning_rate": 8e-05, "loss": 1.634, "step": 9021 }, { "epoch": 0.9750351237436508, "grad_norm": 0.4931080937385559, "learning_rate": 8e-05, "loss": 1.6563, "step": 9022 }, { "epoch": 0.9751431968010374, "grad_norm": 0.4822391867637634, "learning_rate": 8e-05, "loss": 1.6208, "step": 9023 }, { "epoch": 0.9752512698584243, "grad_norm": 0.45389702916145325, "learning_rate": 8e-05, "loss": 1.7577, "step": 9024 }, { "epoch": 0.9753593429158111, "grad_norm": 0.4582240581512451, "learning_rate": 8e-05, "loss": 1.6248, "step": 9025 }, { "epoch": 0.9754674159731979, "grad_norm": 0.517681896686554, "learning_rate": 8e-05, "loss": 1.8352, "step": 9026 }, { "epoch": 0.9755754890305847, "grad_norm": 0.45513346791267395, "learning_rate": 8e-05, "loss": 1.7087, "step": 9027 }, { "epoch": 0.9756835620879715, "grad_norm": 0.4638882279396057, "learning_rate": 8e-05, "loss": 1.6034, "step": 9028 }, { "epoch": 0.9757916351453583, "grad_norm": 0.42928412556648254, "learning_rate": 8e-05, "loss": 1.6643, "step": 9029 }, { "epoch": 0.9758997082027451, "grad_norm": 0.4978237450122833, "learning_rate": 8e-05, "loss": 1.5743, "step": 9030 }, { "epoch": 0.9760077812601319, "grad_norm": 0.4255375862121582, "learning_rate": 8e-05, "loss": 1.5055, "step": 9031 }, { "epoch": 0.9761158543175187, "grad_norm": 0.44315779209136963, "learning_rate": 8e-05, "loss": 1.6037, "step": 9032 }, { "epoch": 0.9762239273749055, "grad_norm": 0.4519886374473572, "learning_rate": 8e-05, "loss": 1.6504, "step": 9033 }, { "epoch": 0.9763320004322922, "grad_norm": 0.46099939942359924, "learning_rate": 8e-05, "loss": 1.9408, "step": 9034 }, { "epoch": 0.976440073489679, "grad_norm": 0.4850521981716156, "learning_rate": 8e-05, "loss": 1.6896, "step": 9035 }, { "epoch": 0.9765481465470658, "grad_norm": 0.43282628059387207, "learning_rate": 8e-05, "loss": 1.5324, "step": 9036 }, { "epoch": 0.9766562196044526, "grad_norm": 0.42202797532081604, "learning_rate": 8e-05, "loss": 1.5511, "step": 9037 }, { "epoch": 0.9767642926618394, "grad_norm": 0.4496881365776062, "learning_rate": 8e-05, "loss": 1.6876, "step": 9038 }, { "epoch": 0.9768723657192262, "grad_norm": 0.4274451732635498, "learning_rate": 8e-05, "loss": 1.6747, "step": 9039 }, { "epoch": 0.976980438776613, "grad_norm": 0.4546601176261902, "learning_rate": 8e-05, "loss": 1.6085, "step": 9040 }, { "epoch": 0.9770885118339998, "grad_norm": 0.4520641565322876, "learning_rate": 8e-05, "loss": 1.6216, "step": 9041 }, { "epoch": 0.9771965848913866, "grad_norm": 0.4838956892490387, "learning_rate": 8e-05, "loss": 1.8191, "step": 9042 }, { "epoch": 0.9773046579487734, "grad_norm": 0.43271589279174805, "learning_rate": 8e-05, "loss": 1.474, "step": 9043 }, { "epoch": 0.9774127310061602, "grad_norm": 0.4460945725440979, "learning_rate": 8e-05, "loss": 1.5973, "step": 9044 }, { "epoch": 0.9775208040635469, "grad_norm": 0.42409825325012207, "learning_rate": 8e-05, "loss": 1.5537, "step": 9045 }, { "epoch": 0.9776288771209337, "grad_norm": 0.47543978691101074, "learning_rate": 8e-05, "loss": 1.4331, "step": 9046 }, { "epoch": 0.9777369501783205, "grad_norm": 0.44449296593666077, "learning_rate": 8e-05, "loss": 1.6666, "step": 9047 }, { "epoch": 0.9778450232357073, "grad_norm": 0.5066308975219727, "learning_rate": 8e-05, "loss": 1.7397, "step": 9048 }, { "epoch": 0.9779530962930941, "grad_norm": 0.4433225095272064, "learning_rate": 8e-05, "loss": 1.6901, "step": 9049 }, { "epoch": 0.9780611693504809, "grad_norm": 0.47958245873451233, "learning_rate": 8e-05, "loss": 1.6358, "step": 9050 }, { "epoch": 0.9781692424078677, "grad_norm": 0.47257936000823975, "learning_rate": 8e-05, "loss": 1.6512, "step": 9051 }, { "epoch": 0.9782773154652545, "grad_norm": 0.5146636366844177, "learning_rate": 8e-05, "loss": 1.7845, "step": 9052 }, { "epoch": 0.9783853885226413, "grad_norm": 0.43844717741012573, "learning_rate": 8e-05, "loss": 1.5674, "step": 9053 }, { "epoch": 0.9784934615800281, "grad_norm": 0.5347436666488647, "learning_rate": 8e-05, "loss": 1.854, "step": 9054 }, { "epoch": 0.9786015346374148, "grad_norm": 0.4546431601047516, "learning_rate": 8e-05, "loss": 1.5778, "step": 9055 }, { "epoch": 0.9787096076948016, "grad_norm": 0.43972712755203247, "learning_rate": 8e-05, "loss": 1.6436, "step": 9056 }, { "epoch": 0.9788176807521884, "grad_norm": 0.4755985736846924, "learning_rate": 8e-05, "loss": 1.6221, "step": 9057 }, { "epoch": 0.9789257538095752, "grad_norm": 0.5375683307647705, "learning_rate": 8e-05, "loss": 1.8102, "step": 9058 }, { "epoch": 0.979033826866962, "grad_norm": 0.48082876205444336, "learning_rate": 8e-05, "loss": 1.6931, "step": 9059 }, { "epoch": 0.9791418999243489, "grad_norm": 0.4435516595840454, "learning_rate": 8e-05, "loss": 1.6856, "step": 9060 }, { "epoch": 0.9792499729817357, "grad_norm": 0.4352591335773468, "learning_rate": 8e-05, "loss": 1.6285, "step": 9061 }, { "epoch": 0.9793580460391225, "grad_norm": 0.46289244294166565, "learning_rate": 8e-05, "loss": 1.6628, "step": 9062 }, { "epoch": 0.9794661190965093, "grad_norm": 0.4339161217212677, "learning_rate": 8e-05, "loss": 1.6948, "step": 9063 }, { "epoch": 0.9795741921538961, "grad_norm": 0.4679482579231262, "learning_rate": 8e-05, "loss": 1.7402, "step": 9064 }, { "epoch": 0.9796822652112829, "grad_norm": 0.4639950096607208, "learning_rate": 8e-05, "loss": 1.6436, "step": 9065 }, { "epoch": 0.9797903382686696, "grad_norm": 0.4516746401786804, "learning_rate": 8e-05, "loss": 1.6379, "step": 9066 }, { "epoch": 0.9798984113260564, "grad_norm": 0.5047026872634888, "learning_rate": 8e-05, "loss": 1.5982, "step": 9067 }, { "epoch": 0.9800064843834432, "grad_norm": 0.424091100692749, "learning_rate": 8e-05, "loss": 1.6306, "step": 9068 }, { "epoch": 0.98011455744083, "grad_norm": 0.4669710397720337, "learning_rate": 8e-05, "loss": 1.8174, "step": 9069 }, { "epoch": 0.9802226304982168, "grad_norm": 0.4289858341217041, "learning_rate": 8e-05, "loss": 1.5729, "step": 9070 }, { "epoch": 0.9803307035556036, "grad_norm": 0.5022605657577515, "learning_rate": 8e-05, "loss": 1.7888, "step": 9071 }, { "epoch": 0.9804387766129904, "grad_norm": 0.4498436152935028, "learning_rate": 8e-05, "loss": 1.7834, "step": 9072 }, { "epoch": 0.9805468496703772, "grad_norm": 0.44242554903030396, "learning_rate": 8e-05, "loss": 1.4822, "step": 9073 }, { "epoch": 0.980654922727764, "grad_norm": 0.451982319355011, "learning_rate": 8e-05, "loss": 1.7172, "step": 9074 }, { "epoch": 0.9807629957851508, "grad_norm": 0.4451686441898346, "learning_rate": 8e-05, "loss": 1.6276, "step": 9075 }, { "epoch": 0.9808710688425376, "grad_norm": 0.4763680696487427, "learning_rate": 8e-05, "loss": 1.6376, "step": 9076 }, { "epoch": 0.9809791418999243, "grad_norm": 0.4979807734489441, "learning_rate": 8e-05, "loss": 1.8123, "step": 9077 }, { "epoch": 0.9810872149573111, "grad_norm": 0.4657883644104004, "learning_rate": 8e-05, "loss": 1.8111, "step": 9078 }, { "epoch": 0.9811952880146979, "grad_norm": 0.46862661838531494, "learning_rate": 8e-05, "loss": 1.5501, "step": 9079 }, { "epoch": 0.9813033610720847, "grad_norm": 0.4528345763683319, "learning_rate": 8e-05, "loss": 1.6632, "step": 9080 }, { "epoch": 0.9814114341294715, "grad_norm": 0.44899633526802063, "learning_rate": 8e-05, "loss": 1.5783, "step": 9081 }, { "epoch": 0.9815195071868583, "grad_norm": 0.512934148311615, "learning_rate": 8e-05, "loss": 1.5517, "step": 9082 }, { "epoch": 0.9816275802442451, "grad_norm": 0.44088172912597656, "learning_rate": 8e-05, "loss": 1.7065, "step": 9083 }, { "epoch": 0.9817356533016319, "grad_norm": 0.4605644643306732, "learning_rate": 8e-05, "loss": 1.4928, "step": 9084 }, { "epoch": 0.9818437263590187, "grad_norm": 0.4591098725795746, "learning_rate": 8e-05, "loss": 1.5534, "step": 9085 }, { "epoch": 0.9819517994164055, "grad_norm": 0.5127196907997131, "learning_rate": 8e-05, "loss": 1.7614, "step": 9086 }, { "epoch": 0.9820598724737923, "grad_norm": 0.4698149561882019, "learning_rate": 8e-05, "loss": 1.7542, "step": 9087 }, { "epoch": 0.982167945531179, "grad_norm": 0.43799877166748047, "learning_rate": 8e-05, "loss": 1.6332, "step": 9088 }, { "epoch": 0.9822760185885658, "grad_norm": 0.451369971036911, "learning_rate": 8e-05, "loss": 1.5865, "step": 9089 }, { "epoch": 0.9823840916459526, "grad_norm": 0.49257132411003113, "learning_rate": 8e-05, "loss": 1.904, "step": 9090 }, { "epoch": 0.9824921647033394, "grad_norm": 0.4817517399787903, "learning_rate": 8e-05, "loss": 1.7574, "step": 9091 }, { "epoch": 0.9826002377607262, "grad_norm": 0.4469441771507263, "learning_rate": 8e-05, "loss": 1.6449, "step": 9092 }, { "epoch": 0.982708310818113, "grad_norm": 0.44638511538505554, "learning_rate": 8e-05, "loss": 1.7029, "step": 9093 }, { "epoch": 0.9828163838754999, "grad_norm": 0.4440476596355438, "learning_rate": 8e-05, "loss": 1.7228, "step": 9094 }, { "epoch": 0.9829244569328867, "grad_norm": 0.4897567331790924, "learning_rate": 8e-05, "loss": 1.7128, "step": 9095 }, { "epoch": 0.9830325299902735, "grad_norm": 0.4599263072013855, "learning_rate": 8e-05, "loss": 1.5989, "step": 9096 }, { "epoch": 0.9831406030476603, "grad_norm": 0.45494699478149414, "learning_rate": 8e-05, "loss": 1.6963, "step": 9097 }, { "epoch": 0.983248676105047, "grad_norm": 0.3918823301792145, "learning_rate": 8e-05, "loss": 1.42, "step": 9098 }, { "epoch": 0.9833567491624338, "grad_norm": 0.5398412942886353, "learning_rate": 8e-05, "loss": 1.7197, "step": 9099 }, { "epoch": 0.9834648222198206, "grad_norm": 0.4580385088920593, "learning_rate": 8e-05, "loss": 1.4966, "step": 9100 }, { "epoch": 0.9835728952772074, "grad_norm": 0.46167895197868347, "learning_rate": 8e-05, "loss": 1.711, "step": 9101 }, { "epoch": 0.9836809683345942, "grad_norm": 0.4857100248336792, "learning_rate": 8e-05, "loss": 1.7633, "step": 9102 }, { "epoch": 0.983789041391981, "grad_norm": 0.4593430161476135, "learning_rate": 8e-05, "loss": 1.7243, "step": 9103 }, { "epoch": 0.9838971144493678, "grad_norm": 0.4855339825153351, "learning_rate": 8e-05, "loss": 1.7218, "step": 9104 }, { "epoch": 0.9840051875067546, "grad_norm": 0.41917097568511963, "learning_rate": 8e-05, "loss": 1.6712, "step": 9105 }, { "epoch": 0.9841132605641414, "grad_norm": 0.45382946729660034, "learning_rate": 8e-05, "loss": 1.7047, "step": 9106 }, { "epoch": 0.9842213336215282, "grad_norm": 0.45571085810661316, "learning_rate": 8e-05, "loss": 1.5627, "step": 9107 }, { "epoch": 0.984329406678915, "grad_norm": 0.45278459787368774, "learning_rate": 8e-05, "loss": 1.6452, "step": 9108 }, { "epoch": 0.9844374797363017, "grad_norm": 0.44884103536605835, "learning_rate": 8e-05, "loss": 1.6441, "step": 9109 }, { "epoch": 0.9845455527936885, "grad_norm": 0.4375130832195282, "learning_rate": 8e-05, "loss": 1.4621, "step": 9110 }, { "epoch": 0.9846536258510753, "grad_norm": 0.42639830708503723, "learning_rate": 8e-05, "loss": 1.5124, "step": 9111 }, { "epoch": 0.9847616989084621, "grad_norm": 0.4799253046512604, "learning_rate": 8e-05, "loss": 1.8005, "step": 9112 }, { "epoch": 0.9848697719658489, "grad_norm": 0.5038084983825684, "learning_rate": 8e-05, "loss": 1.6602, "step": 9113 }, { "epoch": 0.9849778450232357, "grad_norm": 0.5309970378875732, "learning_rate": 8e-05, "loss": 1.6181, "step": 9114 }, { "epoch": 0.9850859180806225, "grad_norm": 0.44439318776130676, "learning_rate": 8e-05, "loss": 1.701, "step": 9115 }, { "epoch": 0.9851939911380093, "grad_norm": 0.46285906434059143, "learning_rate": 8e-05, "loss": 1.5842, "step": 9116 }, { "epoch": 0.9853020641953961, "grad_norm": 0.5051897168159485, "learning_rate": 8e-05, "loss": 1.7597, "step": 9117 }, { "epoch": 0.9854101372527829, "grad_norm": 0.4873689115047455, "learning_rate": 8e-05, "loss": 1.6558, "step": 9118 }, { "epoch": 0.9855182103101697, "grad_norm": 0.4717651307582855, "learning_rate": 8e-05, "loss": 1.6295, "step": 9119 }, { "epoch": 0.9856262833675564, "grad_norm": 0.43424156308174133, "learning_rate": 8e-05, "loss": 1.6617, "step": 9120 }, { "epoch": 0.9857343564249432, "grad_norm": 0.423420786857605, "learning_rate": 8e-05, "loss": 1.6635, "step": 9121 }, { "epoch": 0.98584242948233, "grad_norm": 0.4747917652130127, "learning_rate": 8e-05, "loss": 1.5213, "step": 9122 }, { "epoch": 0.9859505025397168, "grad_norm": 0.4868948757648468, "learning_rate": 8e-05, "loss": 1.5937, "step": 9123 }, { "epoch": 0.9860585755971036, "grad_norm": 0.4352966547012329, "learning_rate": 8e-05, "loss": 1.6891, "step": 9124 }, { "epoch": 0.9861666486544904, "grad_norm": 0.5059141516685486, "learning_rate": 8e-05, "loss": 1.7807, "step": 9125 }, { "epoch": 0.9862747217118772, "grad_norm": 0.45042476058006287, "learning_rate": 8e-05, "loss": 1.7793, "step": 9126 }, { "epoch": 0.986382794769264, "grad_norm": 0.41691359877586365, "learning_rate": 8e-05, "loss": 1.5388, "step": 9127 }, { "epoch": 0.9864908678266509, "grad_norm": 0.5495074987411499, "learning_rate": 8e-05, "loss": 1.6421, "step": 9128 }, { "epoch": 0.9865989408840377, "grad_norm": 0.46722954511642456, "learning_rate": 8e-05, "loss": 1.7546, "step": 9129 }, { "epoch": 0.9867070139414243, "grad_norm": 0.4500752389431, "learning_rate": 8e-05, "loss": 1.6112, "step": 9130 }, { "epoch": 0.9868150869988112, "grad_norm": 0.48936983942985535, "learning_rate": 8e-05, "loss": 1.7596, "step": 9131 }, { "epoch": 0.986923160056198, "grad_norm": 0.485246866941452, "learning_rate": 8e-05, "loss": 1.7659, "step": 9132 }, { "epoch": 0.9870312331135848, "grad_norm": 0.4882313311100006, "learning_rate": 8e-05, "loss": 1.7353, "step": 9133 }, { "epoch": 0.9871393061709716, "grad_norm": 0.48416632413864136, "learning_rate": 8e-05, "loss": 1.7662, "step": 9134 }, { "epoch": 0.9872473792283584, "grad_norm": 0.4613438844680786, "learning_rate": 8e-05, "loss": 1.6805, "step": 9135 }, { "epoch": 0.9873554522857452, "grad_norm": 0.4423687756061554, "learning_rate": 8e-05, "loss": 1.6089, "step": 9136 }, { "epoch": 0.987463525343132, "grad_norm": 0.4881035089492798, "learning_rate": 8e-05, "loss": 1.6458, "step": 9137 }, { "epoch": 0.9875715984005188, "grad_norm": 0.48897626996040344, "learning_rate": 8e-05, "loss": 1.8208, "step": 9138 }, { "epoch": 0.9876796714579056, "grad_norm": 0.4726599454879761, "learning_rate": 8e-05, "loss": 1.6337, "step": 9139 }, { "epoch": 0.9877877445152924, "grad_norm": 0.41692808270454407, "learning_rate": 8e-05, "loss": 1.6414, "step": 9140 }, { "epoch": 0.9878958175726791, "grad_norm": 0.4627240002155304, "learning_rate": 8e-05, "loss": 1.6101, "step": 9141 }, { "epoch": 0.9880038906300659, "grad_norm": 0.5254765152931213, "learning_rate": 8e-05, "loss": 1.7248, "step": 9142 }, { "epoch": 0.9881119636874527, "grad_norm": 0.4261229932308197, "learning_rate": 8e-05, "loss": 1.4004, "step": 9143 }, { "epoch": 0.9882200367448395, "grad_norm": 0.46002399921417236, "learning_rate": 8e-05, "loss": 1.5969, "step": 9144 }, { "epoch": 0.9883281098022263, "grad_norm": 0.5090503096580505, "learning_rate": 8e-05, "loss": 1.6938, "step": 9145 }, { "epoch": 0.9884361828596131, "grad_norm": 0.48529744148254395, "learning_rate": 8e-05, "loss": 1.7785, "step": 9146 }, { "epoch": 0.9885442559169999, "grad_norm": 0.46542462706565857, "learning_rate": 8e-05, "loss": 1.6719, "step": 9147 }, { "epoch": 0.9886523289743867, "grad_norm": 0.4717956781387329, "learning_rate": 8e-05, "loss": 1.7128, "step": 9148 }, { "epoch": 0.9887604020317735, "grad_norm": 0.44755828380584717, "learning_rate": 8e-05, "loss": 1.6444, "step": 9149 }, { "epoch": 0.9888684750891603, "grad_norm": 0.4280109107494354, "learning_rate": 8e-05, "loss": 1.5556, "step": 9150 }, { "epoch": 0.9889765481465471, "grad_norm": 0.4594331979751587, "learning_rate": 8e-05, "loss": 1.8124, "step": 9151 }, { "epoch": 0.9890846212039338, "grad_norm": 0.4246916174888611, "learning_rate": 8e-05, "loss": 1.4797, "step": 9152 }, { "epoch": 0.9891926942613206, "grad_norm": 0.459890753030777, "learning_rate": 8e-05, "loss": 1.696, "step": 9153 }, { "epoch": 0.9893007673187074, "grad_norm": 0.4672141373157501, "learning_rate": 8e-05, "loss": 1.7543, "step": 9154 }, { "epoch": 0.9894088403760942, "grad_norm": 0.47254469990730286, "learning_rate": 8e-05, "loss": 1.7324, "step": 9155 }, { "epoch": 0.989516913433481, "grad_norm": 0.44136494398117065, "learning_rate": 8e-05, "loss": 1.6558, "step": 9156 }, { "epoch": 0.9896249864908678, "grad_norm": 0.4772123098373413, "learning_rate": 8e-05, "loss": 1.7082, "step": 9157 }, { "epoch": 0.9897330595482546, "grad_norm": 0.4541446566581726, "learning_rate": 8e-05, "loss": 1.7661, "step": 9158 }, { "epoch": 0.9898411326056414, "grad_norm": 0.45148441195487976, "learning_rate": 8e-05, "loss": 1.6863, "step": 9159 }, { "epoch": 0.9899492056630282, "grad_norm": 0.4532565176486969, "learning_rate": 8e-05, "loss": 1.6976, "step": 9160 }, { "epoch": 0.990057278720415, "grad_norm": 0.4661872386932373, "learning_rate": 8e-05, "loss": 1.5991, "step": 9161 }, { "epoch": 0.9901653517778017, "grad_norm": 0.44635337591171265, "learning_rate": 8e-05, "loss": 1.6906, "step": 9162 }, { "epoch": 0.9902734248351885, "grad_norm": 0.43400996923446655, "learning_rate": 8e-05, "loss": 1.4397, "step": 9163 }, { "epoch": 0.9903814978925753, "grad_norm": 0.42255210876464844, "learning_rate": 8e-05, "loss": 1.4358, "step": 9164 }, { "epoch": 0.9904895709499622, "grad_norm": 0.4283839166164398, "learning_rate": 8e-05, "loss": 1.4711, "step": 9165 }, { "epoch": 0.990597644007349, "grad_norm": 0.49900686740875244, "learning_rate": 8e-05, "loss": 1.6505, "step": 9166 }, { "epoch": 0.9907057170647358, "grad_norm": 0.4602460563182831, "learning_rate": 8e-05, "loss": 1.686, "step": 9167 }, { "epoch": 0.9908137901221226, "grad_norm": 0.5092211961746216, "learning_rate": 8e-05, "loss": 1.6865, "step": 9168 }, { "epoch": 0.9909218631795094, "grad_norm": 0.4422646760940552, "learning_rate": 8e-05, "loss": 1.4823, "step": 9169 }, { "epoch": 0.9910299362368962, "grad_norm": 0.45360276103019714, "learning_rate": 8e-05, "loss": 1.7258, "step": 9170 }, { "epoch": 0.991138009294283, "grad_norm": 0.4654817581176758, "learning_rate": 8e-05, "loss": 1.6937, "step": 9171 }, { "epoch": 0.9912460823516698, "grad_norm": 0.4393139183521271, "learning_rate": 8e-05, "loss": 1.6076, "step": 9172 }, { "epoch": 0.9913541554090565, "grad_norm": 0.41935181617736816, "learning_rate": 8e-05, "loss": 1.6259, "step": 9173 }, { "epoch": 0.9914622284664433, "grad_norm": 0.4481689929962158, "learning_rate": 8e-05, "loss": 1.6416, "step": 9174 }, { "epoch": 0.9915703015238301, "grad_norm": 0.45338547229766846, "learning_rate": 8e-05, "loss": 1.6362, "step": 9175 }, { "epoch": 0.9916783745812169, "grad_norm": 0.4331096112728119, "learning_rate": 8e-05, "loss": 1.6377, "step": 9176 }, { "epoch": 0.9917864476386037, "grad_norm": 0.4484296441078186, "learning_rate": 8e-05, "loss": 1.6662, "step": 9177 }, { "epoch": 0.9918945206959905, "grad_norm": 0.45766589045524597, "learning_rate": 8e-05, "loss": 1.666, "step": 9178 }, { "epoch": 0.9920025937533773, "grad_norm": 0.42300060391426086, "learning_rate": 8e-05, "loss": 1.5822, "step": 9179 }, { "epoch": 0.9921106668107641, "grad_norm": 0.4256218373775482, "learning_rate": 8e-05, "loss": 1.6184, "step": 9180 }, { "epoch": 0.9922187398681509, "grad_norm": 0.4330958425998688, "learning_rate": 8e-05, "loss": 1.5575, "step": 9181 }, { "epoch": 0.9923268129255377, "grad_norm": 0.45908311009407043, "learning_rate": 8e-05, "loss": 1.7112, "step": 9182 }, { "epoch": 0.9924348859829245, "grad_norm": 0.43104299902915955, "learning_rate": 8e-05, "loss": 1.5571, "step": 9183 }, { "epoch": 0.9925429590403112, "grad_norm": 0.4916972219944, "learning_rate": 8e-05, "loss": 1.5188, "step": 9184 }, { "epoch": 0.992651032097698, "grad_norm": 0.4527086913585663, "learning_rate": 8e-05, "loss": 1.6151, "step": 9185 }, { "epoch": 0.9927591051550848, "grad_norm": 0.5073388814926147, "learning_rate": 8e-05, "loss": 1.8597, "step": 9186 }, { "epoch": 0.9928671782124716, "grad_norm": 0.42115679383277893, "learning_rate": 8e-05, "loss": 1.4187, "step": 9187 }, { "epoch": 0.9929752512698584, "grad_norm": 0.4684169590473175, "learning_rate": 8e-05, "loss": 1.5913, "step": 9188 }, { "epoch": 0.9930833243272452, "grad_norm": 0.4595702588558197, "learning_rate": 8e-05, "loss": 1.4991, "step": 9189 }, { "epoch": 0.993191397384632, "grad_norm": 0.4632265567779541, "learning_rate": 8e-05, "loss": 1.7059, "step": 9190 }, { "epoch": 0.9932994704420188, "grad_norm": 0.47467130422592163, "learning_rate": 8e-05, "loss": 1.6511, "step": 9191 }, { "epoch": 0.9934075434994056, "grad_norm": 0.5125325918197632, "learning_rate": 8e-05, "loss": 1.8051, "step": 9192 }, { "epoch": 0.9935156165567924, "grad_norm": 0.44430482387542725, "learning_rate": 8e-05, "loss": 1.592, "step": 9193 }, { "epoch": 0.9936236896141792, "grad_norm": 0.4456523060798645, "learning_rate": 8e-05, "loss": 1.664, "step": 9194 }, { "epoch": 0.9937317626715659, "grad_norm": 0.4480484127998352, "learning_rate": 8e-05, "loss": 1.6304, "step": 9195 }, { "epoch": 0.9938398357289527, "grad_norm": 0.46802711486816406, "learning_rate": 8e-05, "loss": 1.8174, "step": 9196 }, { "epoch": 0.9939479087863395, "grad_norm": 0.4862932562828064, "learning_rate": 8e-05, "loss": 1.6121, "step": 9197 }, { "epoch": 0.9940559818437263, "grad_norm": 0.4672648012638092, "learning_rate": 8e-05, "loss": 1.5202, "step": 9198 }, { "epoch": 0.9941640549011131, "grad_norm": 0.4451300799846649, "learning_rate": 8e-05, "loss": 1.5931, "step": 9199 }, { "epoch": 0.9942721279585, "grad_norm": 0.41263478994369507, "learning_rate": 8e-05, "loss": 1.6608, "step": 9200 }, { "epoch": 0.9943802010158868, "grad_norm": 0.4722957909107208, "learning_rate": 8e-05, "loss": 1.6532, "step": 9201 }, { "epoch": 0.9944882740732736, "grad_norm": 0.4696202278137207, "learning_rate": 8e-05, "loss": 1.6653, "step": 9202 }, { "epoch": 0.9945963471306604, "grad_norm": 0.4521428942680359, "learning_rate": 8e-05, "loss": 1.6597, "step": 9203 }, { "epoch": 0.9947044201880472, "grad_norm": 0.47508397698402405, "learning_rate": 8e-05, "loss": 1.8461, "step": 9204 }, { "epoch": 0.9948124932454339, "grad_norm": 0.47416961193084717, "learning_rate": 8e-05, "loss": 1.7595, "step": 9205 }, { "epoch": 0.9949205663028207, "grad_norm": 0.5354858636856079, "learning_rate": 8e-05, "loss": 1.7125, "step": 9206 }, { "epoch": 0.9950286393602075, "grad_norm": 0.4399183690547943, "learning_rate": 8e-05, "loss": 1.6457, "step": 9207 }, { "epoch": 0.9951367124175943, "grad_norm": 0.47901374101638794, "learning_rate": 8e-05, "loss": 1.5762, "step": 9208 }, { "epoch": 0.9952447854749811, "grad_norm": 0.47055935859680176, "learning_rate": 8e-05, "loss": 1.6741, "step": 9209 }, { "epoch": 0.9953528585323679, "grad_norm": 0.47451862692832947, "learning_rate": 8e-05, "loss": 1.7335, "step": 9210 }, { "epoch": 0.9954609315897547, "grad_norm": 0.4943249523639679, "learning_rate": 8e-05, "loss": 1.8231, "step": 9211 }, { "epoch": 0.9955690046471415, "grad_norm": 0.4680063724517822, "learning_rate": 8e-05, "loss": 1.6456, "step": 9212 }, { "epoch": 0.9956770777045283, "grad_norm": 0.469836950302124, "learning_rate": 8e-05, "loss": 1.6736, "step": 9213 }, { "epoch": 0.9957851507619151, "grad_norm": 0.48316875100135803, "learning_rate": 8e-05, "loss": 1.7233, "step": 9214 }, { "epoch": 0.9958932238193019, "grad_norm": 0.43895143270492554, "learning_rate": 8e-05, "loss": 1.6068, "step": 9215 }, { "epoch": 0.9960012968766886, "grad_norm": 0.42922255396842957, "learning_rate": 8e-05, "loss": 1.5461, "step": 9216 }, { "epoch": 0.9961093699340754, "grad_norm": 0.46065011620521545, "learning_rate": 8e-05, "loss": 1.6097, "step": 9217 }, { "epoch": 0.9962174429914622, "grad_norm": 0.45429161190986633, "learning_rate": 8e-05, "loss": 1.7098, "step": 9218 }, { "epoch": 0.996325516048849, "grad_norm": 0.5142361521720886, "learning_rate": 8e-05, "loss": 1.7118, "step": 9219 }, { "epoch": 0.9964335891062358, "grad_norm": 0.46596646308898926, "learning_rate": 8e-05, "loss": 1.6876, "step": 9220 }, { "epoch": 0.9965416621636226, "grad_norm": 0.42252659797668457, "learning_rate": 8e-05, "loss": 1.6251, "step": 9221 }, { "epoch": 0.9966497352210094, "grad_norm": 0.5193818807601929, "learning_rate": 8e-05, "loss": 1.6588, "step": 9222 }, { "epoch": 0.9967578082783962, "grad_norm": 0.460664302110672, "learning_rate": 8e-05, "loss": 1.7745, "step": 9223 }, { "epoch": 0.996865881335783, "grad_norm": 0.4626879394054413, "learning_rate": 8e-05, "loss": 1.7019, "step": 9224 }, { "epoch": 0.9969739543931698, "grad_norm": 0.4612959921360016, "learning_rate": 8e-05, "loss": 1.6481, "step": 9225 }, { "epoch": 0.9970820274505566, "grad_norm": 0.44423267245292664, "learning_rate": 8e-05, "loss": 1.6262, "step": 9226 }, { "epoch": 0.9971901005079433, "grad_norm": 0.450556218624115, "learning_rate": 8e-05, "loss": 1.6841, "step": 9227 }, { "epoch": 0.9972981735653301, "grad_norm": 0.459572434425354, "learning_rate": 8e-05, "loss": 1.7127, "step": 9228 }, { "epoch": 0.9974062466227169, "grad_norm": 0.4656917452812195, "learning_rate": 8e-05, "loss": 1.7803, "step": 9229 }, { "epoch": 0.9975143196801037, "grad_norm": 0.4746525287628174, "learning_rate": 8e-05, "loss": 1.6735, "step": 9230 }, { "epoch": 0.9976223927374905, "grad_norm": 0.4440767765045166, "learning_rate": 8e-05, "loss": 1.5297, "step": 9231 }, { "epoch": 0.9977304657948773, "grad_norm": 0.5532853007316589, "learning_rate": 8e-05, "loss": 1.5783, "step": 9232 }, { "epoch": 0.9978385388522641, "grad_norm": 0.45228826999664307, "learning_rate": 8e-05, "loss": 1.6453, "step": 9233 }, { "epoch": 0.997946611909651, "grad_norm": 0.46665480732917786, "learning_rate": 8e-05, "loss": 1.682, "step": 9234 }, { "epoch": 0.9980546849670378, "grad_norm": 0.4928026497364044, "learning_rate": 8e-05, "loss": 1.8134, "step": 9235 }, { "epoch": 0.9981627580244246, "grad_norm": 0.4779892563819885, "learning_rate": 8e-05, "loss": 1.772, "step": 9236 }, { "epoch": 0.9982708310818112, "grad_norm": 0.43793052434921265, "learning_rate": 8e-05, "loss": 1.5688, "step": 9237 }, { "epoch": 0.998378904139198, "grad_norm": 0.486675888299942, "learning_rate": 8e-05, "loss": 1.7747, "step": 9238 }, { "epoch": 0.9984869771965849, "grad_norm": 0.4608488082885742, "learning_rate": 8e-05, "loss": 1.6852, "step": 9239 }, { "epoch": 0.9985950502539717, "grad_norm": 0.46458548307418823, "learning_rate": 8e-05, "loss": 1.7944, "step": 9240 }, { "epoch": 0.9987031233113585, "grad_norm": 0.44192221760749817, "learning_rate": 8e-05, "loss": 1.7121, "step": 9241 }, { "epoch": 0.9988111963687453, "grad_norm": 0.540556013584137, "learning_rate": 8e-05, "loss": 1.8504, "step": 9242 }, { "epoch": 0.9989192694261321, "grad_norm": 0.4775199890136719, "learning_rate": 8e-05, "loss": 1.8522, "step": 9243 }, { "epoch": 0.9990273424835189, "grad_norm": 0.47129446268081665, "learning_rate": 8e-05, "loss": 1.6402, "step": 9244 }, { "epoch": 0.9991354155409057, "grad_norm": 0.4375014007091522, "learning_rate": 8e-05, "loss": 1.6606, "step": 9245 }, { "epoch": 0.9992434885982925, "grad_norm": 0.5035049319267273, "learning_rate": 8e-05, "loss": 1.7296, "step": 9246 }, { "epoch": 0.9993515616556793, "grad_norm": 0.48161736130714417, "learning_rate": 8e-05, "loss": 1.4644, "step": 9247 }, { "epoch": 0.999459634713066, "grad_norm": 0.5007826089859009, "learning_rate": 8e-05, "loss": 1.6732, "step": 9248 }, { "epoch": 0.9995677077704528, "grad_norm": 0.5096982717514038, "learning_rate": 8e-05, "loss": 1.9094, "step": 9249 }, { "epoch": 0.9996757808278396, "grad_norm": 0.4293680191040039, "learning_rate": 8e-05, "loss": 1.5894, "step": 9250 }, { "epoch": 0.9997838538852264, "grad_norm": 0.44377461075782776, "learning_rate": 8e-05, "loss": 1.6623, "step": 9251 }, { "epoch": 0.9998919269426132, "grad_norm": 0.46063360571861267, "learning_rate": 8e-05, "loss": 1.7092, "step": 9252 }, { "epoch": 1.0, "grad_norm": 0.43652763962745667, "learning_rate": 8e-05, "loss": 1.7208, "step": 9253 }, { "epoch": 1.0, "step": 9253, "total_flos": 1.3805669427758236e+19, "train_loss": 0.7757354148797547, "train_runtime": 74869.5284, "train_samples_per_second": 0.494, "train_steps_per_second": 0.124 } ], "logging_steps": 1.0, "max_steps": 9253, "num_input_tokens_seen": 0, "num_train_epochs": 1, "save_steps": 1000, "stateful_callbacks": { "TrainerControl": { "args": { "should_epoch_stop": false, "should_evaluate": false, "should_log": false, "should_save": true, "should_training_stop": true }, "attributes": {} } }, "total_flos": 1.3805669427758236e+19, "train_batch_size": 2, "trial_name": null, "trial_params": null }