{
  "best_metric": null,
  "best_model_checkpoint": null,
  "epoch": 0.8503401360544217,
  "eval_steps": 42,
  "global_step": 500,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0017006802721088435,
      "eval_loss": 3.1163904666900635,
      "eval_runtime": 34.6465,
      "eval_samples_per_second": 28.574,
      "eval_steps_per_second": 3.579,
      "step": 1
    },
    {
      "epoch": 0.00510204081632653,
      "grad_norm": 4.4349870681762695,
      "learning_rate": 3e-05,
      "loss": 3.1357,
      "step": 3
    },
    {
      "epoch": 0.01020408163265306,
      "grad_norm": 2.734741687774658,
      "learning_rate": 6e-05,
      "loss": 2.8293,
      "step": 6
    },
    {
      "epoch": 0.015306122448979591,
      "grad_norm": 3.007612943649292,
      "learning_rate": 9e-05,
      "loss": 2.3471,
      "step": 9
    },
    {
      "epoch": 0.02040816326530612,
      "grad_norm": 2.0711071491241455,
      "learning_rate": 9.999588943391597e-05,
      "loss": 2.2856,
      "step": 12
    },
    {
      "epoch": 0.025510204081632654,
      "grad_norm": 1.5911531448364258,
      "learning_rate": 9.99743108100344e-05,
      "loss": 2.0994,
      "step": 15
    },
    {
      "epoch": 0.030612244897959183,
      "grad_norm": 1.4815939664840698,
      "learning_rate": 9.993424445916923e-05,
      "loss": 2.1158,
      "step": 18
    },
    {
      "epoch": 0.03571428571428571,
      "grad_norm": 1.3731520175933838,
      "learning_rate": 9.987570520365104e-05,
      "loss": 1.9549,
      "step": 21
    },
    {
      "epoch": 0.04081632653061224,
      "grad_norm": 1.3747628927230835,
      "learning_rate": 9.979871469976196e-05,
      "loss": 1.9406,
      "step": 24
    },
    {
      "epoch": 0.04591836734693878,
      "grad_norm": 1.5050859451293945,
      "learning_rate": 9.970330142972401e-05,
      "loss": 1.8567,
      "step": 27
    },
    {
      "epoch": 0.05102040816326531,
      "grad_norm": 1.4547606706619263,
      "learning_rate": 9.95895006911623e-05,
      "loss": 1.8046,
      "step": 30
    },
    {
      "epoch": 0.05612244897959184,
      "grad_norm": 1.4084906578063965,
      "learning_rate": 9.945735458404681e-05,
      "loss": 1.7448,
      "step": 33
    },
    {
      "epoch": 0.061224489795918366,
      "grad_norm": 1.402061939239502,
      "learning_rate": 9.930691199511775e-05,
      "loss": 1.6982,
      "step": 36
    },
    {
      "epoch": 0.0663265306122449,
      "grad_norm": 1.4800869226455688,
      "learning_rate": 9.91382285798002e-05,
      "loss": 1.8587,
      "step": 39
    },
    {
      "epoch": 0.07142857142857142,
      "grad_norm": 1.3802098035812378,
      "learning_rate": 9.895136674161465e-05,
      "loss": 1.7339,
      "step": 42
    },
    {
      "epoch": 0.07142857142857142,
      "eval_loss": 1.721724033355713,
      "eval_runtime": 35.1374,
      "eval_samples_per_second": 28.175,
      "eval_steps_per_second": 3.529,
      "step": 42
    },
    {
      "epoch": 0.07653061224489796,
      "grad_norm": 1.3696004152297974,
      "learning_rate": 9.874639560909117e-05,
      "loss": 1.7873,
      "step": 45
    },
    {
      "epoch": 0.08163265306122448,
      "grad_norm": 1.4703681468963623,
      "learning_rate": 9.852339101019574e-05,
      "loss": 1.6507,
      "step": 48
    },
    {
      "epoch": 0.08673469387755102,
      "grad_norm": 1.4257506132125854,
      "learning_rate": 9.828243544427796e-05,
      "loss": 1.7397,
      "step": 51
    },
    {
      "epoch": 0.09183673469387756,
      "grad_norm": 1.434550166130066,
      "learning_rate": 9.802361805155097e-05,
      "loss": 1.6817,
      "step": 54
    },
    {
      "epoch": 0.09693877551020408,
      "grad_norm": 1.4836714267730713,
      "learning_rate": 9.774703458011453e-05,
      "loss": 1.6046,
      "step": 57
    },
    {
      "epoch": 0.10204081632653061,
      "grad_norm": 1.401002287864685,
      "learning_rate": 9.745278735053343e-05,
      "loss": 1.7368,
      "step": 60
    },
    {
      "epoch": 0.10714285714285714,
      "grad_norm": 1.5878403186798096,
      "learning_rate": 9.714098521798465e-05,
      "loss": 1.623,
      "step": 63
    },
    {
      "epoch": 0.11224489795918367,
      "grad_norm": 1.8310233354568481,
      "learning_rate": 9.681174353198687e-05,
      "loss": 1.608,
      "step": 66
    },
    {
      "epoch": 0.11734693877551021,
      "grad_norm": 1.4681365489959717,
      "learning_rate": 9.64651840937276e-05,
      "loss": 1.6437,
      "step": 69
    },
    {
      "epoch": 0.12244897959183673,
      "grad_norm": 1.3459882736206055,
      "learning_rate": 9.610143511100354e-05,
      "loss": 1.5577,
      "step": 72
    },
    {
      "epoch": 0.12755102040816327,
      "grad_norm": 1.3524380922317505,
      "learning_rate": 9.572063115079063e-05,
      "loss": 1.5852,
      "step": 75
    },
    {
      "epoch": 0.1326530612244898,
      "grad_norm": 1.301050066947937,
      "learning_rate": 9.53229130894619e-05,
      "loss": 1.5688,
      "step": 78
    },
    {
      "epoch": 0.1377551020408163,
      "grad_norm": 1.4519400596618652,
      "learning_rate": 9.490842806067095e-05,
      "loss": 1.6067,
      "step": 81
    },
    {
      "epoch": 0.14285714285714285,
      "grad_norm": 1.2706557512283325,
      "learning_rate": 9.44773294009206e-05,
      "loss": 1.4983,
      "step": 84
    },
    {
      "epoch": 0.14285714285714285,
      "eval_loss": 1.5768018960952759,
      "eval_runtime": 35.1674,
      "eval_samples_per_second": 28.151,
      "eval_steps_per_second": 3.526,
      "step": 84
    },
    {
      "epoch": 0.14795918367346939,
      "grad_norm": 1.302090048789978,
      "learning_rate": 9.40297765928369e-05,
      "loss": 1.523,
      "step": 87
    },
    {
      "epoch": 0.15306122448979592,
      "grad_norm": 1.5333181619644165,
      "learning_rate": 9.356593520616948e-05,
      "loss": 1.6106,
      "step": 90
    },
    {
      "epoch": 0.15816326530612246,
      "grad_norm": 1.360306739807129,
      "learning_rate": 9.308597683653975e-05,
      "loss": 1.5755,
      "step": 93
    },
    {
      "epoch": 0.16326530612244897,
      "grad_norm": 1.4102718830108643,
      "learning_rate": 9.259007904196023e-05,
      "loss": 1.622,
      "step": 96
    },
    {
      "epoch": 0.1683673469387755,
      "grad_norm": 1.19460928440094,
      "learning_rate": 9.207842527714767e-05,
      "loss": 1.5597,
      "step": 99
    },
    {
      "epoch": 0.17346938775510204,
      "grad_norm": 1.374776005744934,
      "learning_rate": 9.155120482565521e-05,
      "loss": 1.6207,
      "step": 102
    },
    {
      "epoch": 0.17857142857142858,
      "grad_norm": 1.337540626525879,
      "learning_rate": 9.10086127298478e-05,
      "loss": 1.5364,
      "step": 105
    },
    {
      "epoch": 0.1836734693877551,
      "grad_norm": 1.4444150924682617,
      "learning_rate": 9.045084971874738e-05,
      "loss": 1.5978,
      "step": 108
    },
    {
      "epoch": 0.18877551020408162,
      "grad_norm": 1.2942811250686646,
      "learning_rate": 8.987812213377424e-05,
      "loss": 1.5385,
      "step": 111
    },
    {
      "epoch": 0.19387755102040816,
      "grad_norm": 1.314165711402893,
      "learning_rate": 8.929064185241213e-05,
      "loss": 1.4721,
      "step": 114
    },
    {
      "epoch": 0.1989795918367347,
      "grad_norm": 1.37766432762146,
      "learning_rate": 8.868862620982534e-05,
      "loss": 1.5167,
      "step": 117
    },
    {
      "epoch": 0.20408163265306123,
      "grad_norm": 1.3688915967941284,
      "learning_rate": 8.807229791845673e-05,
      "loss": 1.516,
      "step": 120
    },
    {
      "epoch": 0.20918367346938777,
      "grad_norm": 1.2788846492767334,
      "learning_rate": 8.744188498563641e-05,
      "loss": 1.4704,
      "step": 123
    },
    {
      "epoch": 0.21428571428571427,
      "grad_norm": 1.3125512599945068,
      "learning_rate": 8.679762062923175e-05,
      "loss": 1.4922,
      "step": 126
    },
    {
      "epoch": 0.21428571428571427,
      "eval_loss": 1.512025237083435,
      "eval_runtime": 35.1759,
      "eval_samples_per_second": 28.144,
      "eval_steps_per_second": 3.525,
      "step": 126
    },
    {
      "epoch": 0.2193877551020408,
      "grad_norm": 1.476131796836853,
      "learning_rate": 8.613974319136958e-05,
      "loss": 1.4581,
      "step": 129
    },
    {
      "epoch": 0.22448979591836735,
      "grad_norm": 1.3448469638824463,
      "learning_rate": 8.54684960502629e-05,
      "loss": 1.5486,
      "step": 132
    },
    {
      "epoch": 0.22959183673469388,
      "grad_norm": 1.416334629058838,
      "learning_rate": 8.478412753017433e-05,
      "loss": 1.5033,
      "step": 135
    },
    {
      "epoch": 0.23469387755102042,
      "grad_norm": 1.442419171333313,
      "learning_rate": 8.408689080954998e-05,
      "loss": 1.5536,
      "step": 138
    },
    {
      "epoch": 0.23979591836734693,
      "grad_norm": 1.3300379514694214,
      "learning_rate": 8.33770438273574e-05,
      "loss": 1.5355,
      "step": 141
    },
    {
      "epoch": 0.24489795918367346,
      "grad_norm": 1.331199288368225,
      "learning_rate": 8.265484918766243e-05,
      "loss": 1.5124,
      "step": 144
    },
    {
      "epoch": 0.25,
      "grad_norm": 1.507230281829834,
      "learning_rate": 8.192057406248028e-05,
      "loss": 1.5505,
      "step": 147
    },
    {
      "epoch": 0.25510204081632654,
      "grad_norm": 1.3450610637664795,
      "learning_rate": 8.117449009293668e-05,
      "loss": 1.5118,
      "step": 150
    },
    {
      "epoch": 0.2602040816326531,
      "grad_norm": 1.3266024589538574,
      "learning_rate": 8.041687328877567e-05,
      "loss": 1.4466,
      "step": 153
    },
    {
      "epoch": 0.2653061224489796,
      "grad_norm": 1.4242159128189087,
      "learning_rate": 7.964800392625129e-05,
      "loss": 1.4573,
      "step": 156
    },
    {
      "epoch": 0.27040816326530615,
      "grad_norm": 1.2786240577697754,
      "learning_rate": 7.886816644444098e-05,
      "loss": 1.4863,
      "step": 159
    },
    {
      "epoch": 0.2755102040816326,
      "grad_norm": 1.326027274131775,
      "learning_rate": 7.807764934001874e-05,
      "loss": 1.4627,
      "step": 162
    },
    {
      "epoch": 0.28061224489795916,
      "grad_norm": 1.366431713104248,
      "learning_rate": 7.727674506052743e-05,
      "loss": 1.3841,
      "step": 165
    },
    {
      "epoch": 0.2857142857142857,
      "grad_norm": 1.2552882432937622,
      "learning_rate": 7.646574989618938e-05,
      "loss": 1.5041,
      "step": 168
    },
    {
      "epoch": 0.2857142857142857,
      "eval_loss": 1.4554890394210815,
      "eval_runtime": 35.2045,
      "eval_samples_per_second": 28.121,
      "eval_steps_per_second": 3.522,
      "step": 168
    },
    {
      "epoch": 0.29081632653061223,
      "grad_norm": 1.4083505868911743,
      "learning_rate": 7.564496387029532e-05,
      "loss": 1.5224,
      "step": 171
    },
    {
      "epoch": 0.29591836734693877,
      "grad_norm": 1.2891498804092407,
      "learning_rate": 7.481469062821252e-05,
      "loss": 1.4262,
      "step": 174
    },
    {
      "epoch": 0.3010204081632653,
      "grad_norm": 1.413690209388733,
      "learning_rate": 7.39752373250527e-05,
      "loss": 1.5159,
      "step": 177
    },
    {
      "epoch": 0.30612244897959184,
      "grad_norm": 1.3442703485488892,
      "learning_rate": 7.312691451204178e-05,
      "loss": 1.4575,
      "step": 180
    },
    {
      "epoch": 0.3112244897959184,
      "grad_norm": 1.348726749420166,
      "learning_rate": 7.227003602163295e-05,
      "loss": 1.5348,
      "step": 183
    },
    {
      "epoch": 0.3163265306122449,
      "grad_norm": 1.338478922843933,
      "learning_rate": 7.14049188514063e-05,
      "loss": 1.4526,
      "step": 186
    },
    {
      "epoch": 0.32142857142857145,
      "grad_norm": 1.3987809419631958,
      "learning_rate": 7.05318830467969e-05,
      "loss": 1.3506,
      "step": 189
    },
    {
      "epoch": 0.32653061224489793,
      "grad_norm": 1.2463765144348145,
      "learning_rate": 6.965125158269619e-05,
      "loss": 1.3693,
      "step": 192
    },
    {
      "epoch": 0.33163265306122447,
      "grad_norm": 1.4282686710357666,
      "learning_rate": 6.876335024396872e-05,
      "loss": 1.5498,
      "step": 195
    },
    {
      "epoch": 0.336734693877551,
      "grad_norm": 1.2759164571762085,
      "learning_rate": 6.786850750493006e-05,
      "loss": 1.4268,
      "step": 198
    },
    {
      "epoch": 0.34183673469387754,
      "grad_norm": 1.4179458618164062,
      "learning_rate": 6.696705440782938e-05,
      "loss": 1.457,
      "step": 201
    },
    {
      "epoch": 0.3469387755102041,
      "grad_norm": 1.480843424797058,
      "learning_rate": 6.605932444038229e-05,
      "loss": 1.4279,
      "step": 204
    },
    {
      "epoch": 0.3520408163265306,
      "grad_norm": 1.2799674272537231,
      "learning_rate": 6.514565341239861e-05,
      "loss": 1.4034,
      "step": 207
    },
    {
      "epoch": 0.35714285714285715,
      "grad_norm": 1.3426426649093628,
      "learning_rate": 6.422637933155162e-05,
      "loss": 1.4177,
      "step": 210
    },
    {
      "epoch": 0.35714285714285715,
      "eval_loss": 1.4206241369247437,
      "eval_runtime": 35.183,
      "eval_samples_per_second": 28.139,
      "eval_steps_per_second": 3.524,
      "step": 210
    },
    {
      "epoch": 0.3622448979591837,
      "grad_norm": 1.2729734182357788,
      "learning_rate": 6.330184227833376e-05,
      "loss": 1.3404,
      "step": 213
    },
    {
      "epoch": 0.3673469387755102,
      "grad_norm": 1.3893803358078003,
      "learning_rate": 6.237238428024572e-05,
      "loss": 1.4067,
      "step": 216
    },
    {
      "epoch": 0.37244897959183676,
      "grad_norm": 1.3455451726913452,
      "learning_rate": 6.143834918526527e-05,
      "loss": 1.3871,
      "step": 219
    },
    {
      "epoch": 0.37755102040816324,
      "grad_norm": 1.4596010446548462,
      "learning_rate": 6.0500082534642464e-05,
      "loss": 1.4465,
      "step": 222
    },
    {
      "epoch": 0.3826530612244898,
      "grad_norm": 1.4347237348556519,
      "learning_rate": 5.955793143506863e-05,
      "loss": 1.4422,
      "step": 225
    },
    {
      "epoch": 0.3877551020408163,
      "grad_norm": 1.2898458242416382,
      "learning_rate": 5.861224443026595e-05,
      "loss": 1.3514,
      "step": 228
    },
    {
      "epoch": 0.39285714285714285,
      "grad_norm": 1.3033252954483032,
      "learning_rate": 5.766337137204579e-05,
      "loss": 1.3825,
      "step": 231
    },
    {
      "epoch": 0.3979591836734694,
      "grad_norm": 1.4201698303222656,
      "learning_rate": 5.6711663290882776e-05,
      "loss": 1.4662,
      "step": 234
    },
    {
      "epoch": 0.4030612244897959,
      "grad_norm": 1.376613974571228,
      "learning_rate": 5.575747226605298e-05,
      "loss": 1.4155,
      "step": 237
    },
    {
      "epoch": 0.40816326530612246,
      "grad_norm": 1.3725097179412842,
      "learning_rate": 5.480115129538409e-05,
      "loss": 1.3242,
      "step": 240
    },
    {
      "epoch": 0.413265306122449,
      "grad_norm": 1.3945027589797974,
      "learning_rate": 5.384305416466584e-05,
      "loss": 1.3763,
      "step": 243
    },
    {
      "epoch": 0.41836734693877553,
      "grad_norm": 1.3332866430282593,
      "learning_rate": 5.288353531676873e-05,
      "loss": 1.4492,
      "step": 246
    },
    {
      "epoch": 0.42346938775510207,
      "grad_norm": 1.439042091369629,
      "learning_rate": 5.192294972051992e-05,
      "loss": 1.4218,
      "step": 249
    },
    {
      "epoch": 0.42857142857142855,
      "grad_norm": 1.4441425800323486,
      "learning_rate": 5.0961652739384356e-05,
      "loss": 1.3857,
      "step": 252
    },
    {
      "epoch": 0.42857142857142855,
      "eval_loss": 1.3904343843460083,
      "eval_runtime": 35.1932,
      "eval_samples_per_second": 28.13,
      "eval_steps_per_second": 3.523,
      "step": 252
    },
    {
      "epoch": 0.4336734693877551,
      "grad_norm": 1.393947958946228,
      "learning_rate": 5e-05,
      "loss": 1.4025,
      "step": 255
    },
    {
      "epoch": 0.4387755102040816,
      "grad_norm": 1.2842843532562256,
      "learning_rate": 4.903834726061565e-05,
      "loss": 1.3806,
      "step": 258
    },
    {
      "epoch": 0.44387755102040816,
      "grad_norm": 1.4686455726623535,
      "learning_rate": 4.807705027948008e-05,
      "loss": 1.3954,
      "step": 261
    },
    {
      "epoch": 0.4489795918367347,
      "grad_norm": 1.3992563486099243,
      "learning_rate": 4.711646468323129e-05,
      "loss": 1.3417,
      "step": 264
    },
    {
      "epoch": 0.45408163265306123,
      "grad_norm": 1.3817628622055054,
      "learning_rate": 4.6156945835334184e-05,
      "loss": 1.4016,
      "step": 267
    },
    {
      "epoch": 0.45918367346938777,
      "grad_norm": 1.4665182828903198,
      "learning_rate": 4.5198848704615914e-05,
      "loss": 1.3706,
      "step": 270
    },
    {
      "epoch": 0.4642857142857143,
      "grad_norm": 1.3997013568878174,
      "learning_rate": 4.424252773394704e-05,
      "loss": 1.4081,
      "step": 273
    },
    {
      "epoch": 0.46938775510204084,
      "grad_norm": 1.4313600063323975,
      "learning_rate": 4.328833670911724e-05,
      "loss": 1.4345,
      "step": 276
    },
    {
      "epoch": 0.4744897959183674,
      "grad_norm": 1.2777554988861084,
      "learning_rate": 4.23366286279542e-05,
      "loss": 1.3368,
      "step": 279
    },
    {
      "epoch": 0.47959183673469385,
      "grad_norm": 1.297582745552063,
      "learning_rate": 4.138775556973406e-05,
      "loss": 1.411,
      "step": 282
    },
    {
      "epoch": 0.4846938775510204,
      "grad_norm": 1.409166693687439,
      "learning_rate": 4.04420685649314e-05,
      "loss": 1.4427,
      "step": 285
    },
    {
      "epoch": 0.4897959183673469,
      "grad_norm": 1.392500400543213,
      "learning_rate": 3.9499917465357534e-05,
      "loss": 1.4439,
      "step": 288
    },
    {
      "epoch": 0.49489795918367346,
      "grad_norm": 1.3200541734695435,
      "learning_rate": 3.856165081473474e-05,
      "loss": 1.39,
      "step": 291
    },
    {
      "epoch": 0.5,
      "grad_norm": 1.4324284791946411,
      "learning_rate": 3.762761571975429e-05,
      "loss": 1.3971,
      "step": 294
    },
    {
      "epoch": 0.5,
      "eval_loss": 1.3641271591186523,
      "eval_runtime": 35.1729,
      "eval_samples_per_second": 28.147,
      "eval_steps_per_second": 3.525,
      "step": 294
    },
    {
      "epoch": 0.5051020408163265,
      "grad_norm": 1.3399404287338257,
      "learning_rate": 3.6698157721666246e-05,
      "loss": 1.3819,
      "step": 297
    },
    {
      "epoch": 0.5102040816326531,
      "grad_norm": 1.3525382280349731,
      "learning_rate": 3.5773620668448384e-05,
      "loss": 1.3289,
      "step": 300
    },
    {
      "epoch": 0.5153061224489796,
      "grad_norm": 1.3362396955490112,
      "learning_rate": 3.48543465876014e-05,
      "loss": 1.3518,
      "step": 303
    },
    {
      "epoch": 0.5204081632653061,
      "grad_norm": 1.4100676774978638,
      "learning_rate": 3.3940675559617724e-05,
      "loss": 1.3504,
      "step": 306
    },
    {
      "epoch": 0.5255102040816326,
      "grad_norm": 1.3914427757263184,
      "learning_rate": 3.303294559217063e-05,
      "loss": 1.4025,
      "step": 309
    },
    {
      "epoch": 0.5306122448979592,
      "grad_norm": 1.3327820301055908,
      "learning_rate": 3.213149249506997e-05,
      "loss": 1.4395,
      "step": 312
    },
    {
      "epoch": 0.5357142857142857,
      "grad_norm": 1.44315505027771,
      "learning_rate": 3.12366497560313e-05,
      "loss": 1.3655,
      "step": 315
    },
    {
      "epoch": 0.5408163265306123,
      "grad_norm": 1.3102085590362549,
      "learning_rate": 3.0348748417303823e-05,
      "loss": 1.3711,
      "step": 318
    },
    {
      "epoch": 0.5459183673469388,
      "grad_norm": 1.31125009059906,
      "learning_rate": 2.9468116953203107e-05,
      "loss": 1.2972,
      "step": 321
    },
    {
      "epoch": 0.5510204081632653,
      "grad_norm": 1.3284008502960205,
      "learning_rate": 2.8595081148593738e-05,
      "loss": 1.3539,
      "step": 324
    },
    {
      "epoch": 0.5561224489795918,
      "grad_norm": 1.274868369102478,
      "learning_rate": 2.772996397836704e-05,
      "loss": 1.3535,
      "step": 327
    },
    {
      "epoch": 0.5612244897959183,
      "grad_norm": 1.3468469381332397,
      "learning_rate": 2.687308548795825e-05,
      "loss": 1.3703,
      "step": 330
    },
    {
      "epoch": 0.5663265306122449,
      "grad_norm": 1.4011824131011963,
      "learning_rate": 2.6024762674947313e-05,
      "loss": 1.3328,
      "step": 333
    },
    {
      "epoch": 0.5714285714285714,
      "grad_norm": 1.2769042253494263,
      "learning_rate": 2.5185309371787513e-05,
      "loss": 1.2754,
      "step": 336
    },
    {
      "epoch": 0.5714285714285714,
      "eval_loss": 1.3424915075302124,
      "eval_runtime": 35.205,
      "eval_samples_per_second": 28.121,
      "eval_steps_per_second": 3.522,
      "step": 336
    },
    {
      "epoch": 0.576530612244898,
      "grad_norm": 1.3609542846679688,
      "learning_rate": 2.43550361297047e-05,
      "loss": 1.3515,
      "step": 339
    },
    {
      "epoch": 0.5816326530612245,
      "grad_norm": 1.3642148971557617,
      "learning_rate": 2.353425010381063e-05,
      "loss": 1.3885,
      "step": 342
    },
    {
      "epoch": 0.5867346938775511,
      "grad_norm": 1.3652516603469849,
      "learning_rate": 2.272325493947257e-05,
      "loss": 1.4246,
      "step": 345
    },
    {
      "epoch": 0.5918367346938775,
      "grad_norm": 1.3225781917572021,
      "learning_rate": 2.192235065998126e-05,
      "loss": 1.3029,
      "step": 348
    },
    {
      "epoch": 0.5969387755102041,
      "grad_norm": 1.3334158658981323,
      "learning_rate": 2.1131833555559037e-05,
      "loss": 1.3672,
      "step": 351
    },
    {
      "epoch": 0.6020408163265306,
      "grad_norm": 1.4105125665664673,
      "learning_rate": 2.0351996073748713e-05,
      "loss": 1.3198,
      "step": 354
    },
    {
      "epoch": 0.6071428571428571,
      "grad_norm": 1.4976215362548828,
      "learning_rate": 1.9583126711224343e-05,
      "loss": 1.3367,
      "step": 357
    },
    {
      "epoch": 0.6122448979591837,
      "grad_norm": 1.3433520793914795,
      "learning_rate": 1.8825509907063327e-05,
      "loss": 1.3276,
      "step": 360
    },
    {
      "epoch": 0.6173469387755102,
      "grad_norm": 1.3432618379592896,
      "learning_rate": 1.807942593751973e-05,
      "loss": 1.366,
      "step": 363
    },
    {
      "epoch": 0.6224489795918368,
      "grad_norm": 1.4558722972869873,
      "learning_rate": 1.7345150812337564e-05,
      "loss": 1.3527,
      "step": 366
    },
    {
      "epoch": 0.6275510204081632,
      "grad_norm": 1.302122712135315,
      "learning_rate": 1.66229561726426e-05,
      "loss": 1.2592,
      "step": 369
    },
    {
      "epoch": 0.6326530612244898,
      "grad_norm": 1.3466747999191284,
      "learning_rate": 1.5913109190450032e-05,
      "loss": 1.3826,
      "step": 372
    },
    {
      "epoch": 0.6377551020408163,
      "grad_norm": 1.2870639562606812,
      "learning_rate": 1.5215872469825682e-05,
      "loss": 1.3761,
      "step": 375
    },
    {
      "epoch": 0.6428571428571429,
      "grad_norm": 1.2575714588165283,
      "learning_rate": 1.4531503949737108e-05,
      "loss": 1.3175,
      "step": 378
    },
    {
      "epoch": 0.6428571428571429,
      "eval_loss": 1.3251529932022095,
      "eval_runtime": 35.1821,
      "eval_samples_per_second": 28.139,
      "eval_steps_per_second": 3.525,
      "step": 378
    },
    {
      "epoch": 0.6479591836734694,
      "grad_norm": 1.3623194694519043,
      "learning_rate": 1.3860256808630428e-05,
      "loss": 1.2629,
      "step": 381
    },
    {
      "epoch": 0.6530612244897959,
      "grad_norm": 1.2920513153076172,
      "learning_rate": 1.3202379370768252e-05,
      "loss": 1.3591,
      "step": 384
    },
    {
      "epoch": 0.6581632653061225,
      "grad_norm": 1.3457062244415283,
      "learning_rate": 1.2558115014363592e-05,
      "loss": 1.3316,
      "step": 387
    },
    {
      "epoch": 0.6632653061224489,
      "grad_norm": 1.3173242807388306,
      "learning_rate": 1.1927702081543279e-05,
      "loss": 1.3149,
      "step": 390
    },
    {
      "epoch": 0.6683673469387755,
      "grad_norm": 1.3412364721298218,
      "learning_rate": 1.1311373790174657e-05,
      "loss": 1.3339,
      "step": 393
    },
    {
      "epoch": 0.673469387755102,
      "grad_norm": 1.3552829027175903,
      "learning_rate": 1.0709358147587884e-05,
      "loss": 1.383,
      "step": 396
    },
    {
      "epoch": 0.6785714285714286,
      "grad_norm": 1.2894299030303955,
      "learning_rate": 1.0121877866225781e-05,
      "loss": 1.3127,
      "step": 399
    },
    {
      "epoch": 0.6836734693877551,
      "grad_norm": 1.3170045614242554,
      "learning_rate": 9.549150281252633e-06,
      "loss": 1.2562,
      "step": 402
    },
    {
      "epoch": 0.6887755102040817,
      "grad_norm": 1.3335039615631104,
      "learning_rate": 8.991387270152201e-06,
      "loss": 1.3243,
      "step": 405
    },
    {
      "epoch": 0.6938775510204082,
      "grad_norm": 1.3929165601730347,
      "learning_rate": 8.448795174344804e-06,
      "loss": 1.3727,
      "step": 408
    },
    {
      "epoch": 0.6989795918367347,
      "grad_norm": 1.3173998594284058,
      "learning_rate": 7.921574722852343e-06,
      "loss": 1.2642,
      "step": 411
    },
    {
      "epoch": 0.7040816326530612,
      "grad_norm": 1.4041547775268555,
      "learning_rate": 7.409920958039795e-06,
      "loss": 1.3234,
      "step": 414
    },
    {
      "epoch": 0.7091836734693877,
      "grad_norm": 1.2738025188446045,
      "learning_rate": 6.9140231634602485e-06,
      "loss": 1.304,
      "step": 417
    },
    {
      "epoch": 0.7142857142857143,
      "grad_norm": 1.36890709400177,
      "learning_rate": 6.43406479383053e-06,
      "loss": 1.2925,
      "step": 420
    },
    {
      "epoch": 0.7142857142857143,
      "eval_loss": 1.3164913654327393,
      "eval_runtime": 35.232,
      "eval_samples_per_second": 28.099,
      "eval_steps_per_second": 3.52,
      "step": 420
    },
    {
      "epoch": 0.7193877551020408,
      "grad_norm": 1.349149227142334,
      "learning_rate": 5.9702234071631e-06,
      "loss": 1.2992,
      "step": 423
    },
    {
      "epoch": 0.7244897959183674,
      "grad_norm": 1.3962756395339966,
      "learning_rate": 5.5226705990794155e-06,
      "loss": 1.3636,
      "step": 426
    },
    {
      "epoch": 0.7295918367346939,
      "grad_norm": 1.3731638193130493,
      "learning_rate": 5.091571939329048e-06,
      "loss": 1.2886,
      "step": 429
    },
    {
      "epoch": 0.7346938775510204,
      "grad_norm": 1.2778152227401733,
      "learning_rate": 4.677086910538092e-06,
      "loss": 1.3175,
      "step": 432
    },
    {
      "epoch": 0.7397959183673469,
      "grad_norm": 1.3971977233886719,
      "learning_rate": 4.279368849209381e-06,
      "loss": 1.3454,
      "step": 435
    },
    {
      "epoch": 0.7448979591836735,
      "grad_norm": 1.2573140859603882,
      "learning_rate": 3.898564888996476e-06,
      "loss": 1.2447,
      "step": 438
    },
    {
      "epoch": 0.75,
      "grad_norm": 1.3133257627487183,
      "learning_rate": 3.534815906272404e-06,
      "loss": 1.2851,
      "step": 441
    },
    {
      "epoch": 0.7551020408163265,
      "grad_norm": 1.426760196685791,
      "learning_rate": 3.18825646801314e-06,
      "loss": 1.3463,
      "step": 444
    },
    {
      "epoch": 0.7602040816326531,
      "grad_norm": 1.2696373462677002,
      "learning_rate": 2.8590147820153513e-06,
      "loss": 1.2552,
      "step": 447
    },
    {
      "epoch": 0.7653061224489796,
      "grad_norm": 1.3841983079910278,
      "learning_rate": 2.547212649466568e-06,
      "loss": 1.36,
      "step": 450
    },
    {
      "epoch": 0.7704081632653061,
      "grad_norm": 1.462028980255127,
      "learning_rate": 2.2529654198854835e-06,
      "loss": 1.3425,
      "step": 453
    },
    {
      "epoch": 0.7755102040816326,
      "grad_norm": 1.3342539072036743,
      "learning_rate": 1.9763819484490355e-06,
      "loss": 1.3417,
      "step": 456
    },
    {
      "epoch": 0.7806122448979592,
      "grad_norm": 1.4027793407440186,
      "learning_rate": 1.7175645557220566e-06,
      "loss": 1.3303,
      "step": 459
    },
    {
      "epoch": 0.7857142857142857,
      "grad_norm": 1.4022960662841797,
      "learning_rate": 1.4766089898042678e-06,
      "loss": 1.3069,
      "step": 462
    },
    {
      "epoch": 0.7857142857142857,
      "eval_loss": 1.3126194477081299,
      "eval_runtime": 35.1877,
      "eval_samples_per_second": 28.135,
      "eval_steps_per_second": 3.524,
      "step": 462
    },
    {
      "epoch": 0.7908163265306123,
      "grad_norm": 1.4359699487686157,
      "learning_rate": 1.2536043909088191e-06,
      "loss": 1.3433,
      "step": 465
    },
    {
      "epoch": 0.7959183673469388,
      "grad_norm": 1.4439855813980103,
      "learning_rate": 1.0486332583853563e-06,
      "loss": 1.3676,
      "step": 468
    },
    {
      "epoch": 0.8010204081632653,
      "grad_norm": 1.370235800743103,
      "learning_rate": 8.617714201998084e-07,
      "loss": 1.2855,
      "step": 471
    },
    {
      "epoch": 0.8061224489795918,
      "grad_norm": 1.2090818881988525,
      "learning_rate": 6.93088004882253e-07,
      "loss": 1.3047,
      "step": 474
    },
    {
      "epoch": 0.8112244897959183,
      "grad_norm": 1.3970741033554077,
      "learning_rate": 5.426454159531913e-07,
      "loss": 1.3344,
      "step": 477
    },
    {
      "epoch": 0.8163265306122449,
      "grad_norm": 1.4051525592803955,
      "learning_rate": 4.104993088376974e-07,
      "loss": 1.3377,
      "step": 480
    },
    {
      "epoch": 0.8214285714285714,
      "grad_norm": 1.4082175493240356,
      "learning_rate": 2.966985702759828e-07,
      "loss": 1.2679,
      "step": 483
    },
    {
      "epoch": 0.826530612244898,
      "grad_norm": 1.3774138689041138,
      "learning_rate": 2.012853002380466e-07,
      "loss": 1.2442,
      "step": 486
    },
    {
      "epoch": 0.8316326530612245,
      "grad_norm": 1.3718559741973877,
      "learning_rate": 1.2429479634897267e-07,
      "loss": 1.3134,
      "step": 489
    },
    {
      "epoch": 0.8367346938775511,
      "grad_norm": 1.3869266510009766,
      "learning_rate": 6.575554083078084e-08,
      "loss": 1.365,
      "step": 492
    },
    {
      "epoch": 0.8418367346938775,
      "grad_norm": 1.3461697101593018,
      "learning_rate": 2.568918996560532e-08,
      "loss": 1.3113,
      "step": 495
    },
    {
      "epoch": 0.8469387755102041,
      "grad_norm": 1.4545960426330566,
      "learning_rate": 4.110566084036816e-09,
      "loss": 1.3021,
      "step": 498
    }
  ],
  "logging_steps": 3,
  "max_steps": 500,
  "num_input_tokens_seen": 0,
  "num_train_epochs": 1,
  "save_steps": 42,
  "stateful_callbacks": {
    "TrainerControl": {
      "args": {
        "should_epoch_stop": false,
        "should_evaluate": false,
        "should_log": false,
        "should_save": true,
        "should_training_stop": true
      },
      "attributes": {}
    }
  },
  "total_flos": 2.02538779410432e+17,
  "train_batch_size": 8,
  "trial_name": null,
  "trial_params": null
}