{
  "best_metric": 4.483307838439941,
  "best_model_checkpoint": "/mmfs1/gscratch/stf/abhinavp/corpus-filtering/outputs/npi-sent-neg/lstm/2/checkpoints/checkpoint-152640",
  "epoch": 1.0250006060157382,
  "eval_steps": 10,
  "global_step": 152640,
  "is_hyper_param_search": false,
  "is_local_process_zero": true,
  "is_world_process_zero": true,
  "log_history": [
    {
      "epoch": 0.0,
      "learning_rate": 4.999998362119627e-05,
      "loss": 10.821,
      "step": 1
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.999161405248948e-05,
      "loss": 7.5575,
      "step": 512
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.998322810497896e-05,
      "loss": 7.0608,
      "step": 1024
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.997484215746844e-05,
      "loss": 6.9947,
      "step": 1536
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.996645620995792e-05,
      "loss": 6.9499,
      "step": 2048
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.99580702624474e-05,
      "loss": 6.9296,
      "step": 2560
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.994968431493688e-05,
      "loss": 6.7676,
      "step": 3072
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.994129836742636e-05,
      "loss": 6.6563,
      "step": 3584
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.993292879871958e-05,
      "loss": 6.5586,
      "step": 4096
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.992454285120906e-05,
      "loss": 6.4769,
      "step": 4608
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.991615690369854e-05,
      "loss": 6.4157,
      "step": 5120
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.990777095618801e-05,
      "loss": 6.3458,
      "step": 5632
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.989938500867749e-05,
      "loss": 6.2756,
      "step": 6144
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.989099906116697e-05,
      "loss": 6.195,
      "step": 6656
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.988261311365645e-05,
      "loss": 6.1472,
      "step": 7168
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.987422716614593e-05,
      "loss": 6.0861,
      "step": 7680
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.986584121863541e-05,
      "loss": 6.0442,
      "step": 8192
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.985745527112489e-05,
      "loss": 5.9999,
      "step": 8704
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.984906932361437e-05,
      "loss": 5.9629,
      "step": 9216
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.984068337610385e-05,
      "loss": 5.9224,
      "step": 9728
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.983231380739706e-05,
      "loss": 5.8906,
      "step": 10240
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9823927859886547e-05,
      "loss": 5.8508,
      "step": 10752
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9815558291179756e-05,
      "loss": 5.8189,
      "step": 11264
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9807172343669236e-05,
      "loss": 5.7878,
      "step": 11776
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9798786396158716e-05,
      "loss": 5.7688,
      "step": 12288
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9790400448648195e-05,
      "loss": 5.7318,
      "step": 12800
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9782014501137675e-05,
      "loss": 5.7079,
      "step": 13312
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9773628553627155e-05,
      "loss": 5.6803,
      "step": 13824
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9765242606116635e-05,
      "loss": 5.6499,
      "step": 14336
    },
    {
      "epoch": 0.0,
      "learning_rate": 4.9756856658606115e-05,
      "loss": 5.6395,
      "step": 14848
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9748470711095595e-05,
      "loss": 5.6144,
      "step": 15360
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9740101142388804e-05,
      "loss": 5.6054,
      "step": 15872
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9731715194878284e-05,
      "loss": 5.568,
      "step": 16384
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9723329247367764e-05,
      "loss": 5.5573,
      "step": 16896
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9714943299857244e-05,
      "loss": 5.5458,
      "step": 17408
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.970657373115046e-05,
      "loss": 5.5288,
      "step": 17920
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.969818778363994e-05,
      "loss": 5.5077,
      "step": 18432
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.968980183612942e-05,
      "loss": 5.473,
      "step": 18944
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.96814158886189e-05,
      "loss": 5.4765,
      "step": 19456
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.967304631991211e-05,
      "loss": 5.4427,
      "step": 19968
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.966466037240159e-05,
      "loss": 5.4448,
      "step": 20480
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.965627442489107e-05,
      "loss": 5.4288,
      "step": 20992
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.964788847738054e-05,
      "loss": 5.4171,
      "step": 21504
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.963951890867376e-05,
      "loss": 5.3984,
      "step": 22016
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.963113296116324e-05,
      "loss": 5.3926,
      "step": 22528
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9622763392456454e-05,
      "loss": 5.3923,
      "step": 23040
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9614377444945934e-05,
      "loss": 5.3753,
      "step": 23552
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9605991497435414e-05,
      "loss": 5.3701,
      "step": 24064
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9597605549924894e-05,
      "loss": 5.3431,
      "step": 24576
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9589219602414374e-05,
      "loss": 5.3277,
      "step": 25088
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.958083365490385e-05,
      "loss": 5.3377,
      "step": 25600
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.957244770739333e-05,
      "loss": 5.3255,
      "step": 26112
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.956406175988281e-05,
      "loss": 5.3166,
      "step": 26624
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9555692191176016e-05,
      "loss": 5.2863,
      "step": 27136
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.954732262246923e-05,
      "loss": 5.3018,
      "step": 27648
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.953893667495871e-05,
      "loss": 5.2756,
      "step": 28160
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.953055072744819e-05,
      "loss": 5.2979,
      "step": 28672
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.952218115874141e-05,
      "loss": 5.2484,
      "step": 29184
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.951379521123089e-05,
      "loss": 5.2606,
      "step": 29696
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.950540926372037e-05,
      "loss": 5.2404,
      "step": 30208
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.949702331620985e-05,
      "loss": 5.2295,
      "step": 30720
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.948863736869932e-05,
      "loss": 5.2302,
      "step": 31232
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.94802514211888e-05,
      "loss": 5.2175,
      "step": 31744
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.947186547367828e-05,
      "loss": 5.1963,
      "step": 32256
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.946347952616776e-05,
      "loss": 5.1971,
      "step": 32768
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.945509357865724e-05,
      "loss": 5.2175,
      "step": 33280
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.944670763114672e-05,
      "loss": 5.19,
      "step": 33792
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.94383216836362e-05,
      "loss": 5.1797,
      "step": 34304
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.942993573612568e-05,
      "loss": 5.1684,
      "step": 34816
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.942156616741889e-05,
      "loss": 5.1537,
      "step": 35328
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.941318021990837e-05,
      "loss": 5.1671,
      "step": 35840
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9404794272397856e-05,
      "loss": 5.1584,
      "step": 36352
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9396408324887336e-05,
      "loss": 5.1525,
      "step": 36864
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9388038756180545e-05,
      "loss": 5.1586,
      "step": 37376
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9379652808670025e-05,
      "loss": 5.1425,
      "step": 37888
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9371266861159505e-05,
      "loss": 5.1386,
      "step": 38400
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9362880913648985e-05,
      "loss": 5.1154,
      "step": 38912
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9354494966138465e-05,
      "loss": 5.1159,
      "step": 39424
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9346109018627945e-05,
      "loss": 5.106,
      "step": 39936
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9337723071117425e-05,
      "loss": 5.0973,
      "step": 40448
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9329337123606905e-05,
      "loss": 5.0972,
      "step": 40960
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9320967554900114e-05,
      "loss": 5.0932,
      "step": 41472
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9312581607389594e-05,
      "loss": 5.0968,
      "step": 41984
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.9304195659879074e-05,
      "loss": 5.0863,
      "step": 42496
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.929582609117229e-05,
      "loss": 5.059,
      "step": 43008
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.928744014366177e-05,
      "loss": 5.0705,
      "step": 43520
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.927907057495498e-05,
      "loss": 5.0711,
      "step": 44032
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.927068462744446e-05,
      "loss": 5.0734,
      "step": 44544
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.926229867993394e-05,
      "loss": 5.0503,
      "step": 45056
    },
    {
      "epoch": 0.01,
      "learning_rate": 4.925391273242342e-05,
      "loss": 5.055,
      "step": 45568
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.92455267849129e-05,
      "loss": 5.0458,
      "step": 46080
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.923714083740238e-05,
      "loss": 5.037,
      "step": 46592
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.922875488989185e-05,
      "loss": 5.0334,
      "step": 47104
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.922036894238133e-05,
      "loss": 5.0283,
      "step": 47616
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.921199937367455e-05,
      "loss": 5.016,
      "step": 48128
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.920362980496776e-05,
      "loss": 5.0234,
      "step": 48640
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.919524385745724e-05,
      "loss": 5.0066,
      "step": 49152
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.918685790994672e-05,
      "loss": 4.999,
      "step": 49664
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.91784719624362e-05,
      "loss": 4.9955,
      "step": 50176
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9170086014925676e-05,
      "loss": 4.9994,
      "step": 50688
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9161700067415156e-05,
      "loss": 4.9934,
      "step": 51200
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9153314119904636e-05,
      "loss": 4.9785,
      "step": 51712
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9144928172394116e-05,
      "loss": 4.9784,
      "step": 52224
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9136542224883596e-05,
      "loss": 4.9672,
      "step": 52736
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9128156277373076e-05,
      "loss": 4.9706,
      "step": 53248
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9119770329862556e-05,
      "loss": 4.9481,
      "step": 53760
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9111400761155765e-05,
      "loss": 4.957,
      "step": 54272
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.9103014813645245e-05,
      "loss": 4.9489,
      "step": 54784
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.909464524493846e-05,
      "loss": 4.9456,
      "step": 55296
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.908625929742794e-05,
      "loss": 4.9504,
      "step": 55808
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.907787334991742e-05,
      "loss": 4.9426,
      "step": 56320
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.90694874024069e-05,
      "loss": 4.9225,
      "step": 56832
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.906110145489638e-05,
      "loss": 4.9308,
      "step": 57344
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.905271550738586e-05,
      "loss": 4.9331,
      "step": 57856
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.904432955987534e-05,
      "loss": 4.9257,
      "step": 58368
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.903594361236482e-05,
      "loss": 4.924,
      "step": 58880
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.90275576648543e-05,
      "loss": 4.9157,
      "step": 59392
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.901917171734378e-05,
      "loss": 4.9126,
      "step": 59904
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.901078576983326e-05,
      "loss": 4.9012,
      "step": 60416
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.900239982232274e-05,
      "loss": 4.9074,
      "step": 60928
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8994046632419686e-05,
      "loss": 4.9032,
      "step": 61440
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8985660684909166e-05,
      "loss": 4.8865,
      "step": 61952
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8977274737398646e-05,
      "loss": 4.8899,
      "step": 62464
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8968888789888125e-05,
      "loss": 4.8856,
      "step": 62976
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8960502842377605e-05,
      "loss": 4.8918,
      "step": 63488
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8952116894867085e-05,
      "loss": 4.8689,
      "step": 64000
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8943730947356565e-05,
      "loss": 4.8749,
      "step": 64512
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.893534499984604e-05,
      "loss": 4.8623,
      "step": 65024
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8926975431139254e-05,
      "loss": 4.8702,
      "step": 65536
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8918589483628734e-05,
      "loss": 4.8596,
      "step": 66048
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8910203536118214e-05,
      "loss": 4.8735,
      "step": 66560
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.890181758860769e-05,
      "loss": 4.8584,
      "step": 67072
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.889343164109717e-05,
      "loss": 4.8533,
      "step": 67584
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.888506207239038e-05,
      "loss": 4.8524,
      "step": 68096
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.88766925036836e-05,
      "loss": 4.8452,
      "step": 68608
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.886830655617308e-05,
      "loss": 4.8535,
      "step": 69120
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.885992060866256e-05,
      "loss": 4.8523,
      "step": 69632
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.885153466115204e-05,
      "loss": 4.8436,
      "step": 70144
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.884314871364151e-05,
      "loss": 4.8381,
      "step": 70656
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.883477914493473e-05,
      "loss": 4.8373,
      "step": 71168
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.882639319742421e-05,
      "loss": 4.8286,
      "step": 71680
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.881800724991369e-05,
      "loss": 4.8156,
      "step": 72192
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.880962130240316e-05,
      "loss": 4.8232,
      "step": 72704
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.880123535489264e-05,
      "loss": 4.8154,
      "step": 73216
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.879286578618586e-05,
      "loss": 4.8066,
      "step": 73728
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.878447983867534e-05,
      "loss": 4.8084,
      "step": 74240
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.877609389116482e-05,
      "loss": 4.805,
      "step": 74752
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.87677079436543e-05,
      "loss": 4.8086,
      "step": 75264
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.875932199614378e-05,
      "loss": 4.8105,
      "step": 75776
    },
    {
      "epoch": 0.02,
      "learning_rate": 4.8750952427436986e-05,
      "loss": 4.7973,
      "step": 76288
    },
    {
      "epoch": 0.03,
      "eval_loss": 4.762572765350342,
      "eval_runtime": 297.8957,
      "eval_samples_per_second": 1280.955,
      "eval_steps_per_second": 40.031,
      "step": 76320
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8742566479926466e-05,
      "loss": 4.786,
      "step": 76800
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8734180532415946e-05,
      "loss": 4.7857,
      "step": 77312
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8725794584905426e-05,
      "loss": 4.7952,
      "step": 77824
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8717408637394906e-05,
      "loss": 4.7881,
      "step": 78336
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8709022689884386e-05,
      "loss": 4.7894,
      "step": 78848
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8700636742373866e-05,
      "loss": 4.773,
      "step": 79360
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8692250794863346e-05,
      "loss": 4.7741,
      "step": 79872
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8683864847352826e-05,
      "loss": 4.7694,
      "step": 80384
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.867549527864604e-05,
      "loss": 4.7722,
      "step": 80896
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.866710933113552e-05,
      "loss": 4.7727,
      "step": 81408
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8658723383625e-05,
      "loss": 4.7685,
      "step": 81920
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.865033743611448e-05,
      "loss": 4.7748,
      "step": 82432
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.864195148860396e-05,
      "loss": 4.7438,
      "step": 82944
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.863356554109344e-05,
      "loss": 4.7558,
      "step": 83456
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.862519597238665e-05,
      "loss": 4.7481,
      "step": 83968
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.861681002487613e-05,
      "loss": 4.7395,
      "step": 84480
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.860842407736561e-05,
      "loss": 4.7537,
      "step": 84992
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.860003812985509e-05,
      "loss": 4.7437,
      "step": 85504
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.859165218234457e-05,
      "loss": 4.7411,
      "step": 86016
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.858326623483405e-05,
      "loss": 4.7615,
      "step": 86528
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.857488028732352e-05,
      "loss": 4.7404,
      "step": 87040
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.856651071861674e-05,
      "loss": 4.7351,
      "step": 87552
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8558124771106226e-05,
      "loss": 4.7345,
      "step": 88064
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.85497388235957e-05,
      "loss": 4.7486,
      "step": 88576
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.854135287608518e-05,
      "loss": 4.7283,
      "step": 89088
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8532983307378395e-05,
      "loss": 4.7241,
      "step": 89600
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.8524597359867875e-05,
      "loss": 4.724,
      "step": 90112
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.851621141235735e-05,
      "loss": 4.7197,
      "step": 90624
    },
    {
      "epoch": 1.0,
      "learning_rate": 4.850782546484683e-05,
      "loss": 4.7195,
      "step": 91136
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8499455896140044e-05,
      "loss": 4.714,
      "step": 91648
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8491069948629524e-05,
      "loss": 4.732,
      "step": 92160
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8482684001119e-05,
      "loss": 4.7123,
      "step": 92672
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.847429805360848e-05,
      "loss": 4.7107,
      "step": 93184
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8465912106097964e-05,
      "loss": 4.7131,
      "step": 93696
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8457526158587444e-05,
      "loss": 4.7115,
      "step": 94208
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8449140211076924e-05,
      "loss": 4.7053,
      "step": 94720
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8440754263566404e-05,
      "loss": 4.6884,
      "step": 95232
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.843238469485961e-05,
      "loss": 4.6928,
      "step": 95744
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.842399874734909e-05,
      "loss": 4.6796,
      "step": 96256
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.841561279983857e-05,
      "loss": 4.6963,
      "step": 96768
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.840722685232805e-05,
      "loss": 4.6843,
      "step": 97280
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.839885728362126e-05,
      "loss": 4.6898,
      "step": 97792
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.839047133611074e-05,
      "loss": 4.6791,
      "step": 98304
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.838208538860022e-05,
      "loss": 4.6859,
      "step": 98816
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.83736994410897e-05,
      "loss": 4.6881,
      "step": 99328
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.836531349357918e-05,
      "loss": 4.6795,
      "step": 99840
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.83569439248724e-05,
      "loss": 4.6843,
      "step": 100352
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8348574356165606e-05,
      "loss": 4.67,
      "step": 100864
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8340188408655086e-05,
      "loss": 4.6638,
      "step": 101376
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8331802461144566e-05,
      "loss": 4.6733,
      "step": 101888
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8323416513634046e-05,
      "loss": 4.6702,
      "step": 102400
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8315030566123526e-05,
      "loss": 4.6752,
      "step": 102912
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8306644618613006e-05,
      "loss": 4.6459,
      "step": 103424
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8298258671102486e-05,
      "loss": 4.6689,
      "step": 103936
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8289889102395695e-05,
      "loss": 4.6461,
      "step": 104448
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8281503154885175e-05,
      "loss": 4.6789,
      "step": 104960
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8273117207374655e-05,
      "loss": 4.6428,
      "step": 105472
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8264731259864135e-05,
      "loss": 4.6589,
      "step": 105984
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8256345312353615e-05,
      "loss": 4.6454,
      "step": 106496
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8247959364843095e-05,
      "loss": 4.6394,
      "step": 107008
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.823957341733258e-05,
      "loss": 4.6442,
      "step": 107520
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.823118746982206e-05,
      "loss": 4.6424,
      "step": 108032
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.822281790111527e-05,
      "loss": 4.6249,
      "step": 108544
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.821443195360475e-05,
      "loss": 4.6311,
      "step": 109056
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.820604600609423e-05,
      "loss": 4.647,
      "step": 109568
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.819766005858371e-05,
      "loss": 4.6327,
      "step": 110080
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8189274111073184e-05,
      "loss": 4.6295,
      "step": 110592
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8180888163562664e-05,
      "loss": 4.6263,
      "step": 111104
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8172502216052144e-05,
      "loss": 4.6163,
      "step": 111616
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.816413264734535e-05,
      "loss": 4.6355,
      "step": 112128
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.815574669983483e-05,
      "loss": 4.6296,
      "step": 112640
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.814737713112805e-05,
      "loss": 4.6277,
      "step": 113152
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8138991183617535e-05,
      "loss": 4.636,
      "step": 113664
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.813060523610701e-05,
      "loss": 4.6314,
      "step": 114176
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.812221928859649e-05,
      "loss": 4.6297,
      "step": 114688
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.811383334108597e-05,
      "loss": 4.6087,
      "step": 115200
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.810544739357545e-05,
      "loss": 4.6192,
      "step": 115712
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.809706144606493e-05,
      "loss": 4.613,
      "step": 116224
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.808867549855441e-05,
      "loss": 4.6025,
      "step": 116736
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.808032230865135e-05,
      "loss": 4.6125,
      "step": 117248
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8071936361140827e-05,
      "loss": 4.6122,
      "step": 117760
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8063550413630307e-05,
      "loss": 4.6203,
      "step": 118272
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.8055164466119786e-05,
      "loss": 4.6134,
      "step": 118784
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.804677851860927e-05,
      "loss": 4.5909,
      "step": 119296
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.803839257109875e-05,
      "loss": 4.6021,
      "step": 119808
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.803000662358823e-05,
      "loss": 4.6103,
      "step": 120320
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.802162067607771e-05,
      "loss": 4.6129,
      "step": 120832
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.801323472856719e-05,
      "loss": 4.6032,
      "step": 121344
    },
    {
      "epoch": 1.01,
      "learning_rate": 4.80048651598604e-05,
      "loss": 4.6065,
      "step": 121856
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.799647921234988e-05,
      "loss": 4.6032,
      "step": 122368
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.798809326483936e-05,
      "loss": 4.5896,
      "step": 122880
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.797972369613257e-05,
      "loss": 4.5972,
      "step": 123392
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.797133774862205e-05,
      "loss": 4.596,
      "step": 123904
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.796295180111153e-05,
      "loss": 4.584,
      "step": 124416
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.795456585360101e-05,
      "loss": 4.6008,
      "step": 124928
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.794617990609049e-05,
      "loss": 4.5824,
      "step": 125440
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.793779395857997e-05,
      "loss": 4.5806,
      "step": 125952
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.792940801106946e-05,
      "loss": 4.5819,
      "step": 126464
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.792102206355894e-05,
      "loss": 4.5887,
      "step": 126976
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.791263611604842e-05,
      "loss": 4.5881,
      "step": 127488
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.790426654734163e-05,
      "loss": 4.5752,
      "step": 128000
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7895880599831107e-05,
      "loss": 4.5744,
      "step": 128512
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7887494652320587e-05,
      "loss": 4.5695,
      "step": 129024
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7879108704810067e-05,
      "loss": 4.5787,
      "step": 129536
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7870722757299546e-05,
      "loss": 4.5557,
      "step": 130048
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7862353188592756e-05,
      "loss": 4.5642,
      "step": 130560
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7853967241082236e-05,
      "loss": 4.5703,
      "step": 131072
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7845581293571715e-05,
      "loss": 4.567,
      "step": 131584
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7837195346061195e-05,
      "loss": 4.5667,
      "step": 132096
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.782882577735441e-05,
      "loss": 4.5656,
      "step": 132608
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.782043982984389e-05,
      "loss": 4.5541,
      "step": 133120
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.781205388233337e-05,
      "loss": 4.5636,
      "step": 133632
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.780368431362658e-05,
      "loss": 4.562,
      "step": 134144
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.779529836611606e-05,
      "loss": 4.5595,
      "step": 134656
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.778691241860554e-05,
      "loss": 4.5644,
      "step": 135168
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.777852647109502e-05,
      "loss": 4.5629,
      "step": 135680
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.777014052358449e-05,
      "loss": 4.555,
      "step": 136192
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.776175457607397e-05,
      "loss": 4.5461,
      "step": 136704
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.775336862856345e-05,
      "loss": 4.5577,
      "step": 137216
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.774498268105293e-05,
      "loss": 4.5524,
      "step": 137728
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.773659673354241e-05,
      "loss": 4.5394,
      "step": 138240
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.772821078603189e-05,
      "loss": 4.5499,
      "step": 138752
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.771984121732511e-05,
      "loss": 4.543,
      "step": 139264
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.771145526981459e-05,
      "loss": 4.5532,
      "step": 139776
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.77030857011078e-05,
      "loss": 4.5355,
      "step": 140288
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.769469975359728e-05,
      "loss": 4.5438,
      "step": 140800
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.768631380608676e-05,
      "loss": 4.5234,
      "step": 141312
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.767792785857624e-05,
      "loss": 4.5458,
      "step": 141824
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.766954191106572e-05,
      "loss": 4.5339,
      "step": 142336
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.76611559635552e-05,
      "loss": 4.5499,
      "step": 142848
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.765277001604468e-05,
      "loss": 4.5401,
      "step": 143360
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.764438406853416e-05,
      "loss": 4.5378,
      "step": 143872
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.763599812102364e-05,
      "loss": 4.529,
      "step": 144384
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.762761217351312e-05,
      "loss": 4.5313,
      "step": 144896
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.76192262260026e-05,
      "loss": 4.5446,
      "step": 145408
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.761084027849208e-05,
      "loss": 4.5466,
      "step": 145920
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.760245433098156e-05,
      "loss": 4.5345,
      "step": 146432
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.759406838347104e-05,
      "loss": 4.5318,
      "step": 146944
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.758568243596052e-05,
      "loss": 4.5297,
      "step": 147456
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.757729648845e-05,
      "loss": 4.5259,
      "step": 147968
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7568926919743206e-05,
      "loss": 4.5207,
      "step": 148480
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.756055735103642e-05,
      "loss": 4.5258,
      "step": 148992
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.75521714035259e-05,
      "loss": 4.5187,
      "step": 149504
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.7543785456015375e-05,
      "loss": 4.5137,
      "step": 150016
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.753541588730859e-05,
      "loss": 4.5184,
      "step": 150528
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.752702993979807e-05,
      "loss": 4.5155,
      "step": 151040
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.751866037109128e-05,
      "loss": 4.5202,
      "step": 151552
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.751027442358077e-05,
      "loss": 4.5202,
      "step": 152064
    },
    {
      "epoch": 1.02,
      "learning_rate": 4.750188847607025e-05,
      "loss": 4.5158,
      "step": 152576
    },
    {
      "epoch": 1.03,
      "eval_loss": 4.483307838439941,
      "eval_runtime": 313.194,
      "eval_samples_per_second": 1218.386,
      "eval_steps_per_second": 38.075,
      "step": 152640
    }
  ],
  "logging_steps": 512,
  "max_steps": 3052726,
  "num_train_epochs": 9223372036854775807,
  "save_steps": 10,
  "total_flos": 6.265510129522253e+16,
  "trial_name": null,
  "trial_params": null
}