CodeRM-8B / trainer_state.json
KAKA22's picture
Upload trainer_state.json with huggingface_hub
8266f30 verified
raw
history blame
170 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.001018848700968,
"eval_steps": 500,
"global_step": 982,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0020376974019358125,
"grad_norm": 5.835855484008789,
"learning_rate": 3.3783783783783786e-08,
"loss": 0.3773,
"step": 1
},
{
"epoch": 0.004075394803871625,
"grad_norm": 6.126637935638428,
"learning_rate": 6.756756756756757e-08,
"loss": 0.359,
"step": 2
},
{
"epoch": 0.006113092205807438,
"grad_norm": 6.190367698669434,
"learning_rate": 1.0135135135135137e-07,
"loss": 0.3575,
"step": 3
},
{
"epoch": 0.00815078960774325,
"grad_norm": 6.040923595428467,
"learning_rate": 1.3513513513513515e-07,
"loss": 0.3513,
"step": 4
},
{
"epoch": 0.010188487009679063,
"grad_norm": 5.735762596130371,
"learning_rate": 1.6891891891891894e-07,
"loss": 0.3365,
"step": 5
},
{
"epoch": 0.012226184411614875,
"grad_norm": 4.77271842956543,
"learning_rate": 2.0270270270270273e-07,
"loss": 0.3494,
"step": 6
},
{
"epoch": 0.014263881813550688,
"grad_norm": 4.999124050140381,
"learning_rate": 2.3648648648648652e-07,
"loss": 0.3416,
"step": 7
},
{
"epoch": 0.0163015792154865,
"grad_norm": 4.210795879364014,
"learning_rate": 2.702702702702703e-07,
"loss": 0.3138,
"step": 8
},
{
"epoch": 0.018339276617422313,
"grad_norm": 2.9208009243011475,
"learning_rate": 3.040540540540541e-07,
"loss": 0.2978,
"step": 9
},
{
"epoch": 0.020376974019358125,
"grad_norm": 2.9749715328216553,
"learning_rate": 3.378378378378379e-07,
"loss": 0.2941,
"step": 10
},
{
"epoch": 0.022414671421293938,
"grad_norm": 2.37031888961792,
"learning_rate": 3.716216216216217e-07,
"loss": 0.2859,
"step": 11
},
{
"epoch": 0.02445236882322975,
"grad_norm": 2.144174098968506,
"learning_rate": 4.0540540540540546e-07,
"loss": 0.2767,
"step": 12
},
{
"epoch": 0.026490066225165563,
"grad_norm": 2.0079538822174072,
"learning_rate": 4.3918918918918923e-07,
"loss": 0.2718,
"step": 13
},
{
"epoch": 0.028527763627101375,
"grad_norm": 2.1613566875457764,
"learning_rate": 4.7297297297297305e-07,
"loss": 0.3009,
"step": 14
},
{
"epoch": 0.030565461029037188,
"grad_norm": 1.8342921733856201,
"learning_rate": 5.067567567567568e-07,
"loss": 0.2576,
"step": 15
},
{
"epoch": 0.032603158430973,
"grad_norm": 1.9354841709136963,
"learning_rate": 5.405405405405406e-07,
"loss": 0.2781,
"step": 16
},
{
"epoch": 0.034640855832908816,
"grad_norm": 1.7937740087509155,
"learning_rate": 5.743243243243245e-07,
"loss": 0.2851,
"step": 17
},
{
"epoch": 0.036678553234844626,
"grad_norm": 1.8419092893600464,
"learning_rate": 6.081081081081082e-07,
"loss": 0.2826,
"step": 18
},
{
"epoch": 0.03871625063678044,
"grad_norm": 1.814770221710205,
"learning_rate": 6.418918918918919e-07,
"loss": 0.2653,
"step": 19
},
{
"epoch": 0.04075394803871625,
"grad_norm": 1.7113629579544067,
"learning_rate": 6.756756756756758e-07,
"loss": 0.2695,
"step": 20
},
{
"epoch": 0.04279164544065207,
"grad_norm": 1.6104000806808472,
"learning_rate": 7.094594594594595e-07,
"loss": 0.2605,
"step": 21
},
{
"epoch": 0.044829342842587876,
"grad_norm": 1.8066229820251465,
"learning_rate": 7.432432432432434e-07,
"loss": 0.2654,
"step": 22
},
{
"epoch": 0.04686704024452369,
"grad_norm": 1.6007901430130005,
"learning_rate": 7.770270270270271e-07,
"loss": 0.2592,
"step": 23
},
{
"epoch": 0.0489047376464595,
"grad_norm": 1.728143334388733,
"learning_rate": 8.108108108108109e-07,
"loss": 0.27,
"step": 24
},
{
"epoch": 0.05094243504839532,
"grad_norm": 1.6227835416793823,
"learning_rate": 8.445945945945947e-07,
"loss": 0.2647,
"step": 25
},
{
"epoch": 0.052980132450331126,
"grad_norm": 1.5776833295822144,
"learning_rate": 8.783783783783785e-07,
"loss": 0.256,
"step": 26
},
{
"epoch": 0.05501782985226694,
"grad_norm": 1.5556491613388062,
"learning_rate": 9.121621621621622e-07,
"loss": 0.2611,
"step": 27
},
{
"epoch": 0.05705552725420275,
"grad_norm": 1.6114894151687622,
"learning_rate": 9.459459459459461e-07,
"loss": 0.2666,
"step": 28
},
{
"epoch": 0.05909322465613857,
"grad_norm": 1.508907675743103,
"learning_rate": 9.797297297297298e-07,
"loss": 0.2513,
"step": 29
},
{
"epoch": 0.061130922058074376,
"grad_norm": 1.6146596670150757,
"learning_rate": 1.0135135135135136e-06,
"loss": 0.2735,
"step": 30
},
{
"epoch": 0.06316861946001019,
"grad_norm": 1.691273808479309,
"learning_rate": 1.0472972972972973e-06,
"loss": 0.2637,
"step": 31
},
{
"epoch": 0.065206316861946,
"grad_norm": 1.5835543870925903,
"learning_rate": 1.0810810810810812e-06,
"loss": 0.2657,
"step": 32
},
{
"epoch": 0.06724401426388181,
"grad_norm": 1.5753782987594604,
"learning_rate": 1.114864864864865e-06,
"loss": 0.2477,
"step": 33
},
{
"epoch": 0.06928171166581763,
"grad_norm": 1.6279054880142212,
"learning_rate": 1.148648648648649e-06,
"loss": 0.2603,
"step": 34
},
{
"epoch": 0.07131940906775344,
"grad_norm": 1.789923906326294,
"learning_rate": 1.1824324324324326e-06,
"loss": 0.272,
"step": 35
},
{
"epoch": 0.07335710646968925,
"grad_norm": 1.9096171855926514,
"learning_rate": 1.2162162162162164e-06,
"loss": 0.2405,
"step": 36
},
{
"epoch": 0.07539480387162506,
"grad_norm": 1.6934822797775269,
"learning_rate": 1.25e-06,
"loss": 0.2649,
"step": 37
},
{
"epoch": 0.07743250127356088,
"grad_norm": 1.5827739238739014,
"learning_rate": 1.2837837837837838e-06,
"loss": 0.2433,
"step": 38
},
{
"epoch": 0.07947019867549669,
"grad_norm": 1.7752106189727783,
"learning_rate": 1.3175675675675676e-06,
"loss": 0.2492,
"step": 39
},
{
"epoch": 0.0815078960774325,
"grad_norm": 1.5755575895309448,
"learning_rate": 1.3513513513513515e-06,
"loss": 0.2483,
"step": 40
},
{
"epoch": 0.08354559347936831,
"grad_norm": 1.5823661088943481,
"learning_rate": 1.3851351351351352e-06,
"loss": 0.2436,
"step": 41
},
{
"epoch": 0.08558329088130413,
"grad_norm": 1.4753117561340332,
"learning_rate": 1.418918918918919e-06,
"loss": 0.2312,
"step": 42
},
{
"epoch": 0.08762098828323994,
"grad_norm": 1.6108685731887817,
"learning_rate": 1.4527027027027027e-06,
"loss": 0.2503,
"step": 43
},
{
"epoch": 0.08965868568517575,
"grad_norm": 1.6036980152130127,
"learning_rate": 1.4864864864864868e-06,
"loss": 0.2409,
"step": 44
},
{
"epoch": 0.09169638308711156,
"grad_norm": 1.6441094875335693,
"learning_rate": 1.5202702702702704e-06,
"loss": 0.2486,
"step": 45
},
{
"epoch": 0.09373408048904738,
"grad_norm": 1.6840177774429321,
"learning_rate": 1.5540540540540541e-06,
"loss": 0.2631,
"step": 46
},
{
"epoch": 0.09577177789098319,
"grad_norm": 1.7543444633483887,
"learning_rate": 1.5878378378378378e-06,
"loss": 0.2632,
"step": 47
},
{
"epoch": 0.097809475292919,
"grad_norm": 1.5591580867767334,
"learning_rate": 1.6216216216216219e-06,
"loss": 0.2382,
"step": 48
},
{
"epoch": 0.09984717269485481,
"grad_norm": 1.5250903367996216,
"learning_rate": 1.6554054054054055e-06,
"loss": 0.2431,
"step": 49
},
{
"epoch": 0.10188487009679063,
"grad_norm": 1.5948268175125122,
"learning_rate": 1.6891891891891894e-06,
"loss": 0.2591,
"step": 50
},
{
"epoch": 0.10392256749872644,
"grad_norm": 2.0706777572631836,
"learning_rate": 1.722972972972973e-06,
"loss": 0.2536,
"step": 51
},
{
"epoch": 0.10596026490066225,
"grad_norm": 1.6050032377243042,
"learning_rate": 1.756756756756757e-06,
"loss": 0.2353,
"step": 52
},
{
"epoch": 0.10799796230259806,
"grad_norm": 1.572332739830017,
"learning_rate": 1.7905405405405408e-06,
"loss": 0.2586,
"step": 53
},
{
"epoch": 0.11003565970453388,
"grad_norm": 1.5016673803329468,
"learning_rate": 1.8243243243243245e-06,
"loss": 0.2506,
"step": 54
},
{
"epoch": 0.11207335710646969,
"grad_norm": 1.5977917909622192,
"learning_rate": 1.8581081081081081e-06,
"loss": 0.2377,
"step": 55
},
{
"epoch": 0.1141110545084055,
"grad_norm": 1.4931672811508179,
"learning_rate": 1.8918918918918922e-06,
"loss": 0.2373,
"step": 56
},
{
"epoch": 0.11614875191034131,
"grad_norm": 1.4913355112075806,
"learning_rate": 1.925675675675676e-06,
"loss": 0.243,
"step": 57
},
{
"epoch": 0.11818644931227713,
"grad_norm": 1.3302721977233887,
"learning_rate": 1.9594594594594595e-06,
"loss": 0.2256,
"step": 58
},
{
"epoch": 0.12022414671421294,
"grad_norm": 1.6652206182479858,
"learning_rate": 1.9932432432432434e-06,
"loss": 0.2444,
"step": 59
},
{
"epoch": 0.12226184411614875,
"grad_norm": 1.701822280883789,
"learning_rate": 2.0270270270270273e-06,
"loss": 0.2311,
"step": 60
},
{
"epoch": 0.12429954151808456,
"grad_norm": 2.096843957901001,
"learning_rate": 2.060810810810811e-06,
"loss": 0.2524,
"step": 61
},
{
"epoch": 0.12633723892002038,
"grad_norm": 1.7165014743804932,
"learning_rate": 2.0945945945945946e-06,
"loss": 0.2514,
"step": 62
},
{
"epoch": 0.12837493632195618,
"grad_norm": 1.5645636320114136,
"learning_rate": 2.1283783783783785e-06,
"loss": 0.2425,
"step": 63
},
{
"epoch": 0.130412633723892,
"grad_norm": 1.4007450342178345,
"learning_rate": 2.1621621621621623e-06,
"loss": 0.2489,
"step": 64
},
{
"epoch": 0.13245033112582782,
"grad_norm": 1.485996127128601,
"learning_rate": 2.195945945945946e-06,
"loss": 0.254,
"step": 65
},
{
"epoch": 0.13448802852776362,
"grad_norm": 1.4894458055496216,
"learning_rate": 2.22972972972973e-06,
"loss": 0.2502,
"step": 66
},
{
"epoch": 0.13652572592969944,
"grad_norm": 1.59180748462677,
"learning_rate": 2.2635135135135135e-06,
"loss": 0.2379,
"step": 67
},
{
"epoch": 0.13856342333163527,
"grad_norm": 1.5528825521469116,
"learning_rate": 2.297297297297298e-06,
"loss": 0.2316,
"step": 68
},
{
"epoch": 0.14060112073357106,
"grad_norm": 1.7990005016326904,
"learning_rate": 2.3310810810810813e-06,
"loss": 0.2553,
"step": 69
},
{
"epoch": 0.14263881813550688,
"grad_norm": 1.5510270595550537,
"learning_rate": 2.364864864864865e-06,
"loss": 0.243,
"step": 70
},
{
"epoch": 0.14467651553744268,
"grad_norm": 1.5694071054458618,
"learning_rate": 2.3986486486486486e-06,
"loss": 0.2471,
"step": 71
},
{
"epoch": 0.1467142129393785,
"grad_norm": 1.4359402656555176,
"learning_rate": 2.432432432432433e-06,
"loss": 0.2324,
"step": 72
},
{
"epoch": 0.14875191034131433,
"grad_norm": 1.372413158416748,
"learning_rate": 2.4662162162162163e-06,
"loss": 0.2455,
"step": 73
},
{
"epoch": 0.15078960774325012,
"grad_norm": 1.4357329607009888,
"learning_rate": 2.5e-06,
"loss": 0.2332,
"step": 74
},
{
"epoch": 0.15282730514518594,
"grad_norm": 1.4806591272354126,
"learning_rate": 2.533783783783784e-06,
"loss": 0.2367,
"step": 75
},
{
"epoch": 0.15486500254712177,
"grad_norm": 1.5909092426300049,
"learning_rate": 2.5675675675675675e-06,
"loss": 0.2351,
"step": 76
},
{
"epoch": 0.15690269994905756,
"grad_norm": 1.3923670053482056,
"learning_rate": 2.601351351351352e-06,
"loss": 0.2413,
"step": 77
},
{
"epoch": 0.15894039735099338,
"grad_norm": 1.4559671878814697,
"learning_rate": 2.6351351351351353e-06,
"loss": 0.2365,
"step": 78
},
{
"epoch": 0.16097809475292918,
"grad_norm": 1.5630332231521606,
"learning_rate": 2.668918918918919e-06,
"loss": 0.2431,
"step": 79
},
{
"epoch": 0.163015792154865,
"grad_norm": 1.783199429512024,
"learning_rate": 2.702702702702703e-06,
"loss": 0.2497,
"step": 80
},
{
"epoch": 0.16505348955680083,
"grad_norm": 1.385953664779663,
"learning_rate": 2.7364864864864865e-06,
"loss": 0.2456,
"step": 81
},
{
"epoch": 0.16709118695873662,
"grad_norm": 1.5928620100021362,
"learning_rate": 2.7702702702702703e-06,
"loss": 0.2356,
"step": 82
},
{
"epoch": 0.16912888436067244,
"grad_norm": 1.6051955223083496,
"learning_rate": 2.8040540540540546e-06,
"loss": 0.2443,
"step": 83
},
{
"epoch": 0.17116658176260827,
"grad_norm": 1.5298365354537964,
"learning_rate": 2.837837837837838e-06,
"loss": 0.2354,
"step": 84
},
{
"epoch": 0.17320427916454406,
"grad_norm": 1.4721895456314087,
"learning_rate": 2.871621621621622e-06,
"loss": 0.2271,
"step": 85
},
{
"epoch": 0.17524197656647988,
"grad_norm": 1.3692090511322021,
"learning_rate": 2.9054054054054054e-06,
"loss": 0.2416,
"step": 86
},
{
"epoch": 0.17727967396841568,
"grad_norm": 1.4523481130599976,
"learning_rate": 2.9391891891891893e-06,
"loss": 0.2271,
"step": 87
},
{
"epoch": 0.1793173713703515,
"grad_norm": 1.5602335929870605,
"learning_rate": 2.9729729729729736e-06,
"loss": 0.2498,
"step": 88
},
{
"epoch": 0.18135506877228733,
"grad_norm": 1.478890061378479,
"learning_rate": 3.006756756756757e-06,
"loss": 0.2373,
"step": 89
},
{
"epoch": 0.18339276617422312,
"grad_norm": 1.4697165489196777,
"learning_rate": 3.040540540540541e-06,
"loss": 0.2522,
"step": 90
},
{
"epoch": 0.18543046357615894,
"grad_norm": 1.5102362632751465,
"learning_rate": 3.0743243243243248e-06,
"loss": 0.2403,
"step": 91
},
{
"epoch": 0.18746816097809477,
"grad_norm": 1.3657342195510864,
"learning_rate": 3.1081081081081082e-06,
"loss": 0.2304,
"step": 92
},
{
"epoch": 0.18950585838003056,
"grad_norm": 1.5056520700454712,
"learning_rate": 3.141891891891892e-06,
"loss": 0.2372,
"step": 93
},
{
"epoch": 0.19154355578196638,
"grad_norm": 1.5441231727600098,
"learning_rate": 3.1756756756756755e-06,
"loss": 0.2456,
"step": 94
},
{
"epoch": 0.19358125318390218,
"grad_norm": 1.472184181213379,
"learning_rate": 3.20945945945946e-06,
"loss": 0.2416,
"step": 95
},
{
"epoch": 0.195618950585838,
"grad_norm": 1.4700815677642822,
"learning_rate": 3.2432432432432437e-06,
"loss": 0.2254,
"step": 96
},
{
"epoch": 0.19765664798777383,
"grad_norm": 1.2121925354003906,
"learning_rate": 3.277027027027027e-06,
"loss": 0.2343,
"step": 97
},
{
"epoch": 0.19969434538970962,
"grad_norm": 1.5309983491897583,
"learning_rate": 3.310810810810811e-06,
"loss": 0.233,
"step": 98
},
{
"epoch": 0.20173204279164544,
"grad_norm": 1.422556757926941,
"learning_rate": 3.3445945945945953e-06,
"loss": 0.254,
"step": 99
},
{
"epoch": 0.20376974019358127,
"grad_norm": 1.3337507247924805,
"learning_rate": 3.3783783783783788e-06,
"loss": 0.2443,
"step": 100
},
{
"epoch": 0.20580743759551706,
"grad_norm": 1.3088750839233398,
"learning_rate": 3.4121621621621626e-06,
"loss": 0.2456,
"step": 101
},
{
"epoch": 0.20784513499745289,
"grad_norm": 1.5006299018859863,
"learning_rate": 3.445945945945946e-06,
"loss": 0.2315,
"step": 102
},
{
"epoch": 0.20988283239938868,
"grad_norm": 1.3762035369873047,
"learning_rate": 3.47972972972973e-06,
"loss": 0.2524,
"step": 103
},
{
"epoch": 0.2119205298013245,
"grad_norm": 1.2947945594787598,
"learning_rate": 3.513513513513514e-06,
"loss": 0.2345,
"step": 104
},
{
"epoch": 0.21395822720326033,
"grad_norm": 1.3257865905761719,
"learning_rate": 3.5472972972972973e-06,
"loss": 0.2275,
"step": 105
},
{
"epoch": 0.21599592460519612,
"grad_norm": 1.3588801622390747,
"learning_rate": 3.5810810810810816e-06,
"loss": 0.2302,
"step": 106
},
{
"epoch": 0.21803362200713194,
"grad_norm": 1.3438785076141357,
"learning_rate": 3.6148648648648655e-06,
"loss": 0.2353,
"step": 107
},
{
"epoch": 0.22007131940906777,
"grad_norm": 1.415419340133667,
"learning_rate": 3.648648648648649e-06,
"loss": 0.2232,
"step": 108
},
{
"epoch": 0.22210901681100356,
"grad_norm": 1.2871397733688354,
"learning_rate": 3.6824324324324328e-06,
"loss": 0.2039,
"step": 109
},
{
"epoch": 0.22414671421293939,
"grad_norm": 1.4281516075134277,
"learning_rate": 3.7162162162162162e-06,
"loss": 0.2476,
"step": 110
},
{
"epoch": 0.22618441161487518,
"grad_norm": 1.3740307092666626,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.2206,
"step": 111
},
{
"epoch": 0.228222109016811,
"grad_norm": 1.5596915483474731,
"learning_rate": 3.7837837837837844e-06,
"loss": 0.2337,
"step": 112
},
{
"epoch": 0.23025980641874683,
"grad_norm": 1.501287817955017,
"learning_rate": 3.817567567567567e-06,
"loss": 0.2348,
"step": 113
},
{
"epoch": 0.23229750382068262,
"grad_norm": 1.5411885976791382,
"learning_rate": 3.851351351351352e-06,
"loss": 0.2466,
"step": 114
},
{
"epoch": 0.23433520122261844,
"grad_norm": 1.5837326049804688,
"learning_rate": 3.885135135135135e-06,
"loss": 0.2329,
"step": 115
},
{
"epoch": 0.23637289862455427,
"grad_norm": 1.222288966178894,
"learning_rate": 3.918918918918919e-06,
"loss": 0.2324,
"step": 116
},
{
"epoch": 0.23841059602649006,
"grad_norm": 1.310904860496521,
"learning_rate": 3.952702702702703e-06,
"loss": 0.2338,
"step": 117
},
{
"epoch": 0.24044829342842589,
"grad_norm": 1.4169098138809204,
"learning_rate": 3.986486486486487e-06,
"loss": 0.2461,
"step": 118
},
{
"epoch": 0.24248599083036168,
"grad_norm": 1.6267861127853394,
"learning_rate": 4.020270270270271e-06,
"loss": 0.2506,
"step": 119
},
{
"epoch": 0.2445236882322975,
"grad_norm": 1.396310567855835,
"learning_rate": 4.0540540540540545e-06,
"loss": 0.2169,
"step": 120
},
{
"epoch": 0.24656138563423333,
"grad_norm": 1.3390737771987915,
"learning_rate": 4.087837837837838e-06,
"loss": 0.2468,
"step": 121
},
{
"epoch": 0.24859908303616912,
"grad_norm": 1.418508768081665,
"learning_rate": 4.121621621621622e-06,
"loss": 0.2592,
"step": 122
},
{
"epoch": 0.25063678043810494,
"grad_norm": 1.4217811822891235,
"learning_rate": 4.155405405405405e-06,
"loss": 0.2277,
"step": 123
},
{
"epoch": 0.25267447784004077,
"grad_norm": 1.351367473602295,
"learning_rate": 4.189189189189189e-06,
"loss": 0.2421,
"step": 124
},
{
"epoch": 0.2547121752419766,
"grad_norm": 1.4876950979232788,
"learning_rate": 4.222972972972974e-06,
"loss": 0.2503,
"step": 125
},
{
"epoch": 0.25674987264391236,
"grad_norm": 1.304235577583313,
"learning_rate": 4.256756756756757e-06,
"loss": 0.2376,
"step": 126
},
{
"epoch": 0.2587875700458482,
"grad_norm": 1.3415497541427612,
"learning_rate": 4.290540540540541e-06,
"loss": 0.2386,
"step": 127
},
{
"epoch": 0.260825267447784,
"grad_norm": 1.6345967054367065,
"learning_rate": 4.324324324324325e-06,
"loss": 0.251,
"step": 128
},
{
"epoch": 0.2628629648497198,
"grad_norm": 1.3598498106002808,
"learning_rate": 4.3581081081081085e-06,
"loss": 0.2568,
"step": 129
},
{
"epoch": 0.26490066225165565,
"grad_norm": 1.2988228797912598,
"learning_rate": 4.391891891891892e-06,
"loss": 0.2423,
"step": 130
},
{
"epoch": 0.2669383596535914,
"grad_norm": 1.4423168897628784,
"learning_rate": 4.4256756756756754e-06,
"loss": 0.2565,
"step": 131
},
{
"epoch": 0.26897605705552724,
"grad_norm": 1.4470850229263306,
"learning_rate": 4.45945945945946e-06,
"loss": 0.254,
"step": 132
},
{
"epoch": 0.27101375445746306,
"grad_norm": 1.3750495910644531,
"learning_rate": 4.493243243243244e-06,
"loss": 0.2438,
"step": 133
},
{
"epoch": 0.2730514518593989,
"grad_norm": 1.2969499826431274,
"learning_rate": 4.527027027027027e-06,
"loss": 0.2519,
"step": 134
},
{
"epoch": 0.2750891492613347,
"grad_norm": 1.3548568487167358,
"learning_rate": 4.560810810810811e-06,
"loss": 0.2407,
"step": 135
},
{
"epoch": 0.27712684666327053,
"grad_norm": 1.3551725149154663,
"learning_rate": 4.594594594594596e-06,
"loss": 0.2432,
"step": 136
},
{
"epoch": 0.2791645440652063,
"grad_norm": 1.3198033571243286,
"learning_rate": 4.628378378378379e-06,
"loss": 0.2349,
"step": 137
},
{
"epoch": 0.2812022414671421,
"grad_norm": 1.344118595123291,
"learning_rate": 4.6621621621621625e-06,
"loss": 0.2374,
"step": 138
},
{
"epoch": 0.28323993886907795,
"grad_norm": 1.4489444494247437,
"learning_rate": 4.695945945945946e-06,
"loss": 0.2452,
"step": 139
},
{
"epoch": 0.28527763627101377,
"grad_norm": 1.2641702890396118,
"learning_rate": 4.72972972972973e-06,
"loss": 0.2341,
"step": 140
},
{
"epoch": 0.2873153336729496,
"grad_norm": 1.227349042892456,
"learning_rate": 4.763513513513514e-06,
"loss": 0.2351,
"step": 141
},
{
"epoch": 0.28935303107488536,
"grad_norm": 1.237866759300232,
"learning_rate": 4.797297297297297e-06,
"loss": 0.2416,
"step": 142
},
{
"epoch": 0.2913907284768212,
"grad_norm": 1.341732144355774,
"learning_rate": 4.831081081081082e-06,
"loss": 0.2273,
"step": 143
},
{
"epoch": 0.293428425878757,
"grad_norm": 1.2197740077972412,
"learning_rate": 4.864864864864866e-06,
"loss": 0.2371,
"step": 144
},
{
"epoch": 0.2954661232806928,
"grad_norm": 1.3351991176605225,
"learning_rate": 4.898648648648649e-06,
"loss": 0.237,
"step": 145
},
{
"epoch": 0.29750382068262865,
"grad_norm": 1.364261507987976,
"learning_rate": 4.932432432432433e-06,
"loss": 0.244,
"step": 146
},
{
"epoch": 0.2995415180845644,
"grad_norm": 1.4714289903640747,
"learning_rate": 4.9662162162162165e-06,
"loss": 0.2566,
"step": 147
},
{
"epoch": 0.30157921548650024,
"grad_norm": 1.2321275472640991,
"learning_rate": 5e-06,
"loss": 0.2305,
"step": 148
},
{
"epoch": 0.30361691288843606,
"grad_norm": 1.257879614830017,
"learning_rate": 4.99622641509434e-06,
"loss": 0.2364,
"step": 149
},
{
"epoch": 0.3056546102903719,
"grad_norm": 1.3090922832489014,
"learning_rate": 4.99245283018868e-06,
"loss": 0.2322,
"step": 150
},
{
"epoch": 0.3076923076923077,
"grad_norm": 1.2299717664718628,
"learning_rate": 4.988679245283019e-06,
"loss": 0.2267,
"step": 151
},
{
"epoch": 0.30973000509424353,
"grad_norm": 1.141762375831604,
"learning_rate": 4.984905660377358e-06,
"loss": 0.2332,
"step": 152
},
{
"epoch": 0.3117677024961793,
"grad_norm": 1.4402216672897339,
"learning_rate": 4.981132075471698e-06,
"loss": 0.2479,
"step": 153
},
{
"epoch": 0.3138053998981151,
"grad_norm": 1.4099055528640747,
"learning_rate": 4.977358490566038e-06,
"loss": 0.2379,
"step": 154
},
{
"epoch": 0.31584309730005095,
"grad_norm": 1.2121434211730957,
"learning_rate": 4.973584905660378e-06,
"loss": 0.2328,
"step": 155
},
{
"epoch": 0.31788079470198677,
"grad_norm": 1.2919939756393433,
"learning_rate": 4.969811320754717e-06,
"loss": 0.242,
"step": 156
},
{
"epoch": 0.3199184921039226,
"grad_norm": 1.4471988677978516,
"learning_rate": 4.966037735849057e-06,
"loss": 0.2504,
"step": 157
},
{
"epoch": 0.32195618950585836,
"grad_norm": 2.552502393722534,
"learning_rate": 4.962264150943397e-06,
"loss": 0.2496,
"step": 158
},
{
"epoch": 0.3239938869077942,
"grad_norm": 1.2150771617889404,
"learning_rate": 4.958490566037736e-06,
"loss": 0.2385,
"step": 159
},
{
"epoch": 0.32603158430973,
"grad_norm": 1.3135796785354614,
"learning_rate": 4.954716981132076e-06,
"loss": 0.2355,
"step": 160
},
{
"epoch": 0.3280692817116658,
"grad_norm": 1.3980915546417236,
"learning_rate": 4.950943396226415e-06,
"loss": 0.2338,
"step": 161
},
{
"epoch": 0.33010697911360165,
"grad_norm": 1.3292484283447266,
"learning_rate": 4.947169811320755e-06,
"loss": 0.243,
"step": 162
},
{
"epoch": 0.3321446765155374,
"grad_norm": 1.5226550102233887,
"learning_rate": 4.943396226415095e-06,
"loss": 0.2486,
"step": 163
},
{
"epoch": 0.33418237391747324,
"grad_norm": 1.2037345170974731,
"learning_rate": 4.939622641509435e-06,
"loss": 0.2528,
"step": 164
},
{
"epoch": 0.33622007131940906,
"grad_norm": 1.3524994850158691,
"learning_rate": 4.935849056603774e-06,
"loss": 0.2739,
"step": 165
},
{
"epoch": 0.3382577687213449,
"grad_norm": 1.1955732107162476,
"learning_rate": 4.932075471698114e-06,
"loss": 0.2257,
"step": 166
},
{
"epoch": 0.3402954661232807,
"grad_norm": 1.273659348487854,
"learning_rate": 4.928301886792453e-06,
"loss": 0.2434,
"step": 167
},
{
"epoch": 0.34233316352521653,
"grad_norm": 1.365476369857788,
"learning_rate": 4.924528301886793e-06,
"loss": 0.2269,
"step": 168
},
{
"epoch": 0.3443708609271523,
"grad_norm": 1.3352711200714111,
"learning_rate": 4.920754716981133e-06,
"loss": 0.2413,
"step": 169
},
{
"epoch": 0.3464085583290881,
"grad_norm": 1.2405195236206055,
"learning_rate": 4.916981132075473e-06,
"loss": 0.2382,
"step": 170
},
{
"epoch": 0.34844625573102395,
"grad_norm": 1.4409878253936768,
"learning_rate": 4.913207547169812e-06,
"loss": 0.2379,
"step": 171
},
{
"epoch": 0.35048395313295977,
"grad_norm": 1.269126534461975,
"learning_rate": 4.909433962264152e-06,
"loss": 0.2366,
"step": 172
},
{
"epoch": 0.3525216505348956,
"grad_norm": 1.1738016605377197,
"learning_rate": 4.905660377358491e-06,
"loss": 0.2311,
"step": 173
},
{
"epoch": 0.35455934793683136,
"grad_norm": 1.3719390630722046,
"learning_rate": 4.9018867924528306e-06,
"loss": 0.2238,
"step": 174
},
{
"epoch": 0.3565970453387672,
"grad_norm": 1.301747441291809,
"learning_rate": 4.8981132075471705e-06,
"loss": 0.2376,
"step": 175
},
{
"epoch": 0.358634742740703,
"grad_norm": 1.2950748205184937,
"learning_rate": 4.8943396226415095e-06,
"loss": 0.2614,
"step": 176
},
{
"epoch": 0.36067244014263883,
"grad_norm": 1.3460197448730469,
"learning_rate": 4.8905660377358495e-06,
"loss": 0.247,
"step": 177
},
{
"epoch": 0.36271013754457465,
"grad_norm": 1.1415988206863403,
"learning_rate": 4.886792452830189e-06,
"loss": 0.236,
"step": 178
},
{
"epoch": 0.3647478349465104,
"grad_norm": 1.2343894243240356,
"learning_rate": 4.8830188679245284e-06,
"loss": 0.2338,
"step": 179
},
{
"epoch": 0.36678553234844624,
"grad_norm": 1.3122376203536987,
"learning_rate": 4.879245283018868e-06,
"loss": 0.2512,
"step": 180
},
{
"epoch": 0.36882322975038206,
"grad_norm": 1.1867709159851074,
"learning_rate": 4.875471698113207e-06,
"loss": 0.2319,
"step": 181
},
{
"epoch": 0.3708609271523179,
"grad_norm": 1.242540955543518,
"learning_rate": 4.871698113207547e-06,
"loss": 0.2391,
"step": 182
},
{
"epoch": 0.3728986245542537,
"grad_norm": 1.277694821357727,
"learning_rate": 4.867924528301887e-06,
"loss": 0.2334,
"step": 183
},
{
"epoch": 0.37493632195618953,
"grad_norm": 1.341858983039856,
"learning_rate": 4.864150943396227e-06,
"loss": 0.2363,
"step": 184
},
{
"epoch": 0.3769740193581253,
"grad_norm": 1.2502440214157104,
"learning_rate": 4.860377358490567e-06,
"loss": 0.2342,
"step": 185
},
{
"epoch": 0.3790117167600611,
"grad_norm": 1.2432913780212402,
"learning_rate": 4.856603773584906e-06,
"loss": 0.2436,
"step": 186
},
{
"epoch": 0.38104941416199695,
"grad_norm": 1.3028502464294434,
"learning_rate": 4.852830188679245e-06,
"loss": 0.2367,
"step": 187
},
{
"epoch": 0.38308711156393277,
"grad_norm": 1.3056414127349854,
"learning_rate": 4.849056603773585e-06,
"loss": 0.2221,
"step": 188
},
{
"epoch": 0.3851248089658686,
"grad_norm": 1.1587262153625488,
"learning_rate": 4.845283018867925e-06,
"loss": 0.2511,
"step": 189
},
{
"epoch": 0.38716250636780436,
"grad_norm": 1.3277629613876343,
"learning_rate": 4.841509433962265e-06,
"loss": 0.2376,
"step": 190
},
{
"epoch": 0.3892002037697402,
"grad_norm": 1.3022247552871704,
"learning_rate": 4.837735849056604e-06,
"loss": 0.2503,
"step": 191
},
{
"epoch": 0.391237901171676,
"grad_norm": 1.1903053522109985,
"learning_rate": 4.833962264150944e-06,
"loss": 0.2376,
"step": 192
},
{
"epoch": 0.39327559857361183,
"grad_norm": 1.3128589391708374,
"learning_rate": 4.830188679245284e-06,
"loss": 0.2393,
"step": 193
},
{
"epoch": 0.39531329597554765,
"grad_norm": 2.1321053504943848,
"learning_rate": 4.826415094339623e-06,
"loss": 0.2214,
"step": 194
},
{
"epoch": 0.3973509933774834,
"grad_norm": 1.255610466003418,
"learning_rate": 4.822641509433963e-06,
"loss": 0.2377,
"step": 195
},
{
"epoch": 0.39938869077941924,
"grad_norm": 1.1986833810806274,
"learning_rate": 4.818867924528302e-06,
"loss": 0.2285,
"step": 196
},
{
"epoch": 0.40142638818135506,
"grad_norm": 1.2554630041122437,
"learning_rate": 4.815094339622642e-06,
"loss": 0.2491,
"step": 197
},
{
"epoch": 0.4034640855832909,
"grad_norm": 1.2297279834747314,
"learning_rate": 4.811320754716982e-06,
"loss": 0.2301,
"step": 198
},
{
"epoch": 0.4055017829852267,
"grad_norm": 1.3254568576812744,
"learning_rate": 4.807547169811322e-06,
"loss": 0.2326,
"step": 199
},
{
"epoch": 0.40753948038716253,
"grad_norm": 1.2830324172973633,
"learning_rate": 4.803773584905661e-06,
"loss": 0.2466,
"step": 200
},
{
"epoch": 0.4095771777890983,
"grad_norm": 1.3986701965332031,
"learning_rate": 4.800000000000001e-06,
"loss": 0.2632,
"step": 201
},
{
"epoch": 0.4116148751910341,
"grad_norm": 1.3166711330413818,
"learning_rate": 4.79622641509434e-06,
"loss": 0.2294,
"step": 202
},
{
"epoch": 0.41365257259296995,
"grad_norm": 1.4140809774398804,
"learning_rate": 4.79245283018868e-06,
"loss": 0.2524,
"step": 203
},
{
"epoch": 0.41569026999490577,
"grad_norm": 1.298222303390503,
"learning_rate": 4.7886792452830195e-06,
"loss": 0.2201,
"step": 204
},
{
"epoch": 0.4177279673968416,
"grad_norm": 1.2514641284942627,
"learning_rate": 4.7849056603773594e-06,
"loss": 0.2458,
"step": 205
},
{
"epoch": 0.41976566479877736,
"grad_norm": 1.1963963508605957,
"learning_rate": 4.7811320754716985e-06,
"loss": 0.2363,
"step": 206
},
{
"epoch": 0.4218033622007132,
"grad_norm": 1.2280910015106201,
"learning_rate": 4.777358490566038e-06,
"loss": 0.249,
"step": 207
},
{
"epoch": 0.423841059602649,
"grad_norm": 1.2325594425201416,
"learning_rate": 4.7735849056603775e-06,
"loss": 0.239,
"step": 208
},
{
"epoch": 0.42587875700458483,
"grad_norm": 1.268089771270752,
"learning_rate": 4.769811320754717e-06,
"loss": 0.2428,
"step": 209
},
{
"epoch": 0.42791645440652065,
"grad_norm": 1.147208571434021,
"learning_rate": 4.766037735849057e-06,
"loss": 0.2027,
"step": 210
},
{
"epoch": 0.4299541518084564,
"grad_norm": 1.293784260749817,
"learning_rate": 4.762264150943396e-06,
"loss": 0.2467,
"step": 211
},
{
"epoch": 0.43199184921039224,
"grad_norm": 1.3141661882400513,
"learning_rate": 4.758490566037736e-06,
"loss": 0.235,
"step": 212
},
{
"epoch": 0.43402954661232807,
"grad_norm": 1.3035210371017456,
"learning_rate": 4.754716981132076e-06,
"loss": 0.2461,
"step": 213
},
{
"epoch": 0.4360672440142639,
"grad_norm": 1.26072359085083,
"learning_rate": 4.750943396226415e-06,
"loss": 0.2483,
"step": 214
},
{
"epoch": 0.4381049414161997,
"grad_norm": 1.3666430711746216,
"learning_rate": 4.747169811320755e-06,
"loss": 0.2343,
"step": 215
},
{
"epoch": 0.44014263881813553,
"grad_norm": 1.1508736610412598,
"learning_rate": 4.743396226415094e-06,
"loss": 0.2456,
"step": 216
},
{
"epoch": 0.4421803362200713,
"grad_norm": 1.292580485343933,
"learning_rate": 4.739622641509434e-06,
"loss": 0.2244,
"step": 217
},
{
"epoch": 0.4442180336220071,
"grad_norm": 1.2299766540527344,
"learning_rate": 4.735849056603774e-06,
"loss": 0.2394,
"step": 218
},
{
"epoch": 0.44625573102394295,
"grad_norm": 1.316611886024475,
"learning_rate": 4.732075471698114e-06,
"loss": 0.2301,
"step": 219
},
{
"epoch": 0.44829342842587877,
"grad_norm": 1.3932688236236572,
"learning_rate": 4.728301886792453e-06,
"loss": 0.2184,
"step": 220
},
{
"epoch": 0.4503311258278146,
"grad_norm": 1.2194689512252808,
"learning_rate": 4.724528301886793e-06,
"loss": 0.2312,
"step": 221
},
{
"epoch": 0.45236882322975036,
"grad_norm": 1.1593202352523804,
"learning_rate": 4.720754716981132e-06,
"loss": 0.2122,
"step": 222
},
{
"epoch": 0.4544065206316862,
"grad_norm": 1.2845839262008667,
"learning_rate": 4.716981132075472e-06,
"loss": 0.2309,
"step": 223
},
{
"epoch": 0.456444218033622,
"grad_norm": 1.28933846950531,
"learning_rate": 4.713207547169812e-06,
"loss": 0.2461,
"step": 224
},
{
"epoch": 0.45848191543555783,
"grad_norm": 1.2130182981491089,
"learning_rate": 4.709433962264151e-06,
"loss": 0.2276,
"step": 225
},
{
"epoch": 0.46051961283749365,
"grad_norm": 1.1695858240127563,
"learning_rate": 4.705660377358491e-06,
"loss": 0.2386,
"step": 226
},
{
"epoch": 0.4625573102394294,
"grad_norm": 1.1642833948135376,
"learning_rate": 4.701886792452831e-06,
"loss": 0.2234,
"step": 227
},
{
"epoch": 0.46459500764136524,
"grad_norm": 1.2203588485717773,
"learning_rate": 4.69811320754717e-06,
"loss": 0.2245,
"step": 228
},
{
"epoch": 0.46663270504330107,
"grad_norm": 1.5108790397644043,
"learning_rate": 4.69433962264151e-06,
"loss": 0.2587,
"step": 229
},
{
"epoch": 0.4686704024452369,
"grad_norm": 2.0064799785614014,
"learning_rate": 4.690566037735849e-06,
"loss": 0.2279,
"step": 230
},
{
"epoch": 0.4707080998471727,
"grad_norm": 1.1252530813217163,
"learning_rate": 4.686792452830189e-06,
"loss": 0.23,
"step": 231
},
{
"epoch": 0.47274579724910853,
"grad_norm": 1.359333872795105,
"learning_rate": 4.683018867924529e-06,
"loss": 0.2444,
"step": 232
},
{
"epoch": 0.4747834946510443,
"grad_norm": 1.1184419393539429,
"learning_rate": 4.6792452830188686e-06,
"loss": 0.2339,
"step": 233
},
{
"epoch": 0.4768211920529801,
"grad_norm": 1.2560921907424927,
"learning_rate": 4.6754716981132085e-06,
"loss": 0.2549,
"step": 234
},
{
"epoch": 0.47885888945491595,
"grad_norm": 1.164919137954712,
"learning_rate": 4.6716981132075476e-06,
"loss": 0.238,
"step": 235
},
{
"epoch": 0.48089658685685177,
"grad_norm": 1.3727017641067505,
"learning_rate": 4.667924528301887e-06,
"loss": 0.2434,
"step": 236
},
{
"epoch": 0.4829342842587876,
"grad_norm": 1.4192495346069336,
"learning_rate": 4.6641509433962265e-06,
"loss": 0.2364,
"step": 237
},
{
"epoch": 0.48497198166072336,
"grad_norm": 1.3315473794937134,
"learning_rate": 4.6603773584905665e-06,
"loss": 0.2166,
"step": 238
},
{
"epoch": 0.4870096790626592,
"grad_norm": 1.140080213546753,
"learning_rate": 4.656603773584906e-06,
"loss": 0.2404,
"step": 239
},
{
"epoch": 0.489047376464595,
"grad_norm": 1.3821226358413696,
"learning_rate": 4.6528301886792454e-06,
"loss": 0.2376,
"step": 240
},
{
"epoch": 0.49108507386653083,
"grad_norm": 1.3087458610534668,
"learning_rate": 4.649056603773585e-06,
"loss": 0.2364,
"step": 241
},
{
"epoch": 0.49312277126846665,
"grad_norm": 1.1473658084869385,
"learning_rate": 4.645283018867925e-06,
"loss": 0.2356,
"step": 242
},
{
"epoch": 0.4951604686704024,
"grad_norm": 1.2189340591430664,
"learning_rate": 4.641509433962264e-06,
"loss": 0.2377,
"step": 243
},
{
"epoch": 0.49719816607233824,
"grad_norm": 1.6314069032669067,
"learning_rate": 4.637735849056604e-06,
"loss": 0.2496,
"step": 244
},
{
"epoch": 0.49923586347427407,
"grad_norm": 1.1271792650222778,
"learning_rate": 4.633962264150943e-06,
"loss": 0.2179,
"step": 245
},
{
"epoch": 0.5012735608762099,
"grad_norm": 1.1837356090545654,
"learning_rate": 4.630188679245283e-06,
"loss": 0.2334,
"step": 246
},
{
"epoch": 0.5033112582781457,
"grad_norm": 1.2714848518371582,
"learning_rate": 4.626415094339623e-06,
"loss": 0.263,
"step": 247
},
{
"epoch": 0.5053489556800815,
"grad_norm": 1.235137701034546,
"learning_rate": 4.622641509433963e-06,
"loss": 0.2551,
"step": 248
},
{
"epoch": 0.5073866530820174,
"grad_norm": 1.067122220993042,
"learning_rate": 4.618867924528302e-06,
"loss": 0.2397,
"step": 249
},
{
"epoch": 0.5094243504839532,
"grad_norm": 1.3224409818649292,
"learning_rate": 4.615094339622642e-06,
"loss": 0.2407,
"step": 250
},
{
"epoch": 0.5114620478858889,
"grad_norm": 1.1779237985610962,
"learning_rate": 4.611320754716981e-06,
"loss": 0.2297,
"step": 251
},
{
"epoch": 0.5134997452878247,
"grad_norm": 1.3744945526123047,
"learning_rate": 4.607547169811321e-06,
"loss": 0.241,
"step": 252
},
{
"epoch": 0.5155374426897605,
"grad_norm": 1.198855996131897,
"learning_rate": 4.603773584905661e-06,
"loss": 0.2352,
"step": 253
},
{
"epoch": 0.5175751400916964,
"grad_norm": 1.117774486541748,
"learning_rate": 4.600000000000001e-06,
"loss": 0.2418,
"step": 254
},
{
"epoch": 0.5196128374936322,
"grad_norm": 1.248888611793518,
"learning_rate": 4.59622641509434e-06,
"loss": 0.2244,
"step": 255
},
{
"epoch": 0.521650534895568,
"grad_norm": 1.2080906629562378,
"learning_rate": 4.59245283018868e-06,
"loss": 0.229,
"step": 256
},
{
"epoch": 0.5236882322975038,
"grad_norm": 1.1990790367126465,
"learning_rate": 4.588679245283019e-06,
"loss": 0.2253,
"step": 257
},
{
"epoch": 0.5257259296994397,
"grad_norm": 1.1284271478652954,
"learning_rate": 4.584905660377359e-06,
"loss": 0.2381,
"step": 258
},
{
"epoch": 0.5277636271013755,
"grad_norm": 1.2414554357528687,
"learning_rate": 4.581132075471699e-06,
"loss": 0.2332,
"step": 259
},
{
"epoch": 0.5298013245033113,
"grad_norm": 1.0976932048797607,
"learning_rate": 4.577358490566038e-06,
"loss": 0.2331,
"step": 260
},
{
"epoch": 0.5318390219052471,
"grad_norm": 1.2075899839401245,
"learning_rate": 4.573584905660378e-06,
"loss": 0.2413,
"step": 261
},
{
"epoch": 0.5338767193071828,
"grad_norm": 1.1429880857467651,
"learning_rate": 4.569811320754718e-06,
"loss": 0.23,
"step": 262
},
{
"epoch": 0.5359144167091187,
"grad_norm": 1.0777602195739746,
"learning_rate": 4.566037735849057e-06,
"loss": 0.2185,
"step": 263
},
{
"epoch": 0.5379521141110545,
"grad_norm": 1.143699288368225,
"learning_rate": 4.562264150943397e-06,
"loss": 0.2324,
"step": 264
},
{
"epoch": 0.5399898115129903,
"grad_norm": 1.3619898557662964,
"learning_rate": 4.558490566037736e-06,
"loss": 0.2484,
"step": 265
},
{
"epoch": 0.5420275089149261,
"grad_norm": 1.2638384103775024,
"learning_rate": 4.554716981132076e-06,
"loss": 0.2349,
"step": 266
},
{
"epoch": 0.544065206316862,
"grad_norm": 1.2247638702392578,
"learning_rate": 4.5509433962264155e-06,
"loss": 0.2372,
"step": 267
},
{
"epoch": 0.5461029037187978,
"grad_norm": 1.2362172603607178,
"learning_rate": 4.547169811320755e-06,
"loss": 0.2333,
"step": 268
},
{
"epoch": 0.5481406011207336,
"grad_norm": 1.2307566404342651,
"learning_rate": 4.543396226415095e-06,
"loss": 0.2411,
"step": 269
},
{
"epoch": 0.5501782985226694,
"grad_norm": 1.2503217458724976,
"learning_rate": 4.539622641509434e-06,
"loss": 0.2459,
"step": 270
},
{
"epoch": 0.5522159959246052,
"grad_norm": 1.2475491762161255,
"learning_rate": 4.5358490566037735e-06,
"loss": 0.2272,
"step": 271
},
{
"epoch": 0.5542536933265411,
"grad_norm": 1.105730414390564,
"learning_rate": 4.532075471698113e-06,
"loss": 0.2481,
"step": 272
},
{
"epoch": 0.5562913907284768,
"grad_norm": 1.275002121925354,
"learning_rate": 4.528301886792453e-06,
"loss": 0.2445,
"step": 273
},
{
"epoch": 0.5583290881304126,
"grad_norm": 1.1774675846099854,
"learning_rate": 4.524528301886793e-06,
"loss": 0.2377,
"step": 274
},
{
"epoch": 0.5603667855323484,
"grad_norm": 1.329745888710022,
"learning_rate": 4.520754716981132e-06,
"loss": 0.2319,
"step": 275
},
{
"epoch": 0.5624044829342842,
"grad_norm": 1.2236435413360596,
"learning_rate": 4.516981132075472e-06,
"loss": 0.2274,
"step": 276
},
{
"epoch": 0.5644421803362201,
"grad_norm": 1.0417534112930298,
"learning_rate": 4.513207547169812e-06,
"loss": 0.2234,
"step": 277
},
{
"epoch": 0.5664798777381559,
"grad_norm": 1.0934056043624878,
"learning_rate": 4.509433962264151e-06,
"loss": 0.2445,
"step": 278
},
{
"epoch": 0.5685175751400917,
"grad_norm": 1.2551244497299194,
"learning_rate": 4.505660377358491e-06,
"loss": 0.2434,
"step": 279
},
{
"epoch": 0.5705552725420275,
"grad_norm": 1.2088017463684082,
"learning_rate": 4.50188679245283e-06,
"loss": 0.2399,
"step": 280
},
{
"epoch": 0.5725929699439634,
"grad_norm": 1.2738829851150513,
"learning_rate": 4.49811320754717e-06,
"loss": 0.2328,
"step": 281
},
{
"epoch": 0.5746306673458992,
"grad_norm": 1.312220811843872,
"learning_rate": 4.49433962264151e-06,
"loss": 0.2334,
"step": 282
},
{
"epoch": 0.5766683647478349,
"grad_norm": 1.1316941976547241,
"learning_rate": 4.49056603773585e-06,
"loss": 0.225,
"step": 283
},
{
"epoch": 0.5787060621497707,
"grad_norm": 1.0500327348709106,
"learning_rate": 4.486792452830189e-06,
"loss": 0.226,
"step": 284
},
{
"epoch": 0.5807437595517065,
"grad_norm": 1.0962241888046265,
"learning_rate": 4.483018867924528e-06,
"loss": 0.2168,
"step": 285
},
{
"epoch": 0.5827814569536424,
"grad_norm": 1.100046992301941,
"learning_rate": 4.479245283018868e-06,
"loss": 0.2267,
"step": 286
},
{
"epoch": 0.5848191543555782,
"grad_norm": 1.2387049198150635,
"learning_rate": 4.475471698113208e-06,
"loss": 0.2509,
"step": 287
},
{
"epoch": 0.586856851757514,
"grad_norm": 1.1974563598632812,
"learning_rate": 4.471698113207548e-06,
"loss": 0.2351,
"step": 288
},
{
"epoch": 0.5888945491594498,
"grad_norm": 1.2102775573730469,
"learning_rate": 4.467924528301888e-06,
"loss": 0.2474,
"step": 289
},
{
"epoch": 0.5909322465613857,
"grad_norm": 1.0824848413467407,
"learning_rate": 4.464150943396227e-06,
"loss": 0.2289,
"step": 290
},
{
"epoch": 0.5929699439633215,
"grad_norm": 1.111902117729187,
"learning_rate": 4.460377358490567e-06,
"loss": 0.2363,
"step": 291
},
{
"epoch": 0.5950076413652573,
"grad_norm": 1.1692800521850586,
"learning_rate": 4.456603773584906e-06,
"loss": 0.2266,
"step": 292
},
{
"epoch": 0.5970453387671931,
"grad_norm": 1.160117506980896,
"learning_rate": 4.452830188679246e-06,
"loss": 0.2351,
"step": 293
},
{
"epoch": 0.5990830361691288,
"grad_norm": 1.1320550441741943,
"learning_rate": 4.4490566037735856e-06,
"loss": 0.2239,
"step": 294
},
{
"epoch": 0.6011207335710647,
"grad_norm": 1.1472080945968628,
"learning_rate": 4.445283018867925e-06,
"loss": 0.2157,
"step": 295
},
{
"epoch": 0.6031584309730005,
"grad_norm": 1.2992992401123047,
"learning_rate": 4.4415094339622646e-06,
"loss": 0.2283,
"step": 296
},
{
"epoch": 0.6051961283749363,
"grad_norm": 1.2557927370071411,
"learning_rate": 4.4377358490566045e-06,
"loss": 0.2339,
"step": 297
},
{
"epoch": 0.6072338257768721,
"grad_norm": 1.0591647624969482,
"learning_rate": 4.4339622641509435e-06,
"loss": 0.2152,
"step": 298
},
{
"epoch": 0.609271523178808,
"grad_norm": 1.0702134370803833,
"learning_rate": 4.4301886792452834e-06,
"loss": 0.206,
"step": 299
},
{
"epoch": 0.6113092205807438,
"grad_norm": 1.2004814147949219,
"learning_rate": 4.4264150943396225e-06,
"loss": 0.2307,
"step": 300
},
{
"epoch": 0.6133469179826796,
"grad_norm": 1.1907483339309692,
"learning_rate": 4.4226415094339624e-06,
"loss": 0.2289,
"step": 301
},
{
"epoch": 0.6153846153846154,
"grad_norm": 1.1154402494430542,
"learning_rate": 4.418867924528302e-06,
"loss": 0.2135,
"step": 302
},
{
"epoch": 0.6174223127865512,
"grad_norm": 1.1816260814666748,
"learning_rate": 4.415094339622642e-06,
"loss": 0.2286,
"step": 303
},
{
"epoch": 0.6194600101884871,
"grad_norm": 1.1875680685043335,
"learning_rate": 4.411320754716981e-06,
"loss": 0.2375,
"step": 304
},
{
"epoch": 0.6214977075904228,
"grad_norm": 1.0976321697235107,
"learning_rate": 4.407547169811321e-06,
"loss": 0.2137,
"step": 305
},
{
"epoch": 0.6235354049923586,
"grad_norm": 1.10517418384552,
"learning_rate": 4.40377358490566e-06,
"loss": 0.219,
"step": 306
},
{
"epoch": 0.6255731023942944,
"grad_norm": 1.1795883178710938,
"learning_rate": 4.4e-06,
"loss": 0.2286,
"step": 307
},
{
"epoch": 0.6276107997962302,
"grad_norm": 1.1178569793701172,
"learning_rate": 4.39622641509434e-06,
"loss": 0.22,
"step": 308
},
{
"epoch": 0.6296484971981661,
"grad_norm": 1.1791189908981323,
"learning_rate": 4.39245283018868e-06,
"loss": 0.2474,
"step": 309
},
{
"epoch": 0.6316861946001019,
"grad_norm": 1.1312475204467773,
"learning_rate": 4.388679245283019e-06,
"loss": 0.2474,
"step": 310
},
{
"epoch": 0.6337238920020377,
"grad_norm": 1.1903657913208008,
"learning_rate": 4.384905660377359e-06,
"loss": 0.2477,
"step": 311
},
{
"epoch": 0.6357615894039735,
"grad_norm": 1.1177330017089844,
"learning_rate": 4.381132075471698e-06,
"loss": 0.2412,
"step": 312
},
{
"epoch": 0.6377992868059094,
"grad_norm": 1.3050440549850464,
"learning_rate": 4.377358490566038e-06,
"loss": 0.2465,
"step": 313
},
{
"epoch": 0.6398369842078452,
"grad_norm": 1.1658434867858887,
"learning_rate": 4.373584905660378e-06,
"loss": 0.2162,
"step": 314
},
{
"epoch": 0.6418746816097809,
"grad_norm": 1.1645337343215942,
"learning_rate": 4.369811320754717e-06,
"loss": 0.2369,
"step": 315
},
{
"epoch": 0.6439123790117167,
"grad_norm": 1.1002851724624634,
"learning_rate": 4.366037735849057e-06,
"loss": 0.1995,
"step": 316
},
{
"epoch": 0.6459500764136525,
"grad_norm": 1.1481510400772095,
"learning_rate": 4.362264150943397e-06,
"loss": 0.2324,
"step": 317
},
{
"epoch": 0.6479877738155884,
"grad_norm": 1.2481803894042969,
"learning_rate": 4.358490566037737e-06,
"loss": 0.2327,
"step": 318
},
{
"epoch": 0.6500254712175242,
"grad_norm": 1.1513328552246094,
"learning_rate": 4.354716981132076e-06,
"loss": 0.2585,
"step": 319
},
{
"epoch": 0.65206316861946,
"grad_norm": 1.1497119665145874,
"learning_rate": 4.350943396226415e-06,
"loss": 0.233,
"step": 320
},
{
"epoch": 0.6541008660213958,
"grad_norm": 1.1609100103378296,
"learning_rate": 4.347169811320755e-06,
"loss": 0.2405,
"step": 321
},
{
"epoch": 0.6561385634233317,
"grad_norm": 1.2081102132797241,
"learning_rate": 4.343396226415095e-06,
"loss": 0.2378,
"step": 322
},
{
"epoch": 0.6581762608252675,
"grad_norm": 1.1453851461410522,
"learning_rate": 4.339622641509435e-06,
"loss": 0.2298,
"step": 323
},
{
"epoch": 0.6602139582272033,
"grad_norm": 1.051963210105896,
"learning_rate": 4.3358490566037745e-06,
"loss": 0.2424,
"step": 324
},
{
"epoch": 0.6622516556291391,
"grad_norm": 1.122475028038025,
"learning_rate": 4.332075471698114e-06,
"loss": 0.2217,
"step": 325
},
{
"epoch": 0.6642893530310748,
"grad_norm": 1.557624101638794,
"learning_rate": 4.3283018867924535e-06,
"loss": 0.2476,
"step": 326
},
{
"epoch": 0.6663270504330107,
"grad_norm": 1.133476972579956,
"learning_rate": 4.324528301886793e-06,
"loss": 0.24,
"step": 327
},
{
"epoch": 0.6683647478349465,
"grad_norm": 1.192134976387024,
"learning_rate": 4.3207547169811325e-06,
"loss": 0.2285,
"step": 328
},
{
"epoch": 0.6704024452368823,
"grad_norm": 1.0376332998275757,
"learning_rate": 4.316981132075472e-06,
"loss": 0.2314,
"step": 329
},
{
"epoch": 0.6724401426388181,
"grad_norm": 1.1142336130142212,
"learning_rate": 4.3132075471698115e-06,
"loss": 0.2213,
"step": 330
},
{
"epoch": 0.674477840040754,
"grad_norm": 1.075834035873413,
"learning_rate": 4.309433962264151e-06,
"loss": 0.2541,
"step": 331
},
{
"epoch": 0.6765155374426898,
"grad_norm": 1.6311166286468506,
"learning_rate": 4.305660377358491e-06,
"loss": 0.2273,
"step": 332
},
{
"epoch": 0.6785532348446256,
"grad_norm": 1.2183853387832642,
"learning_rate": 4.30188679245283e-06,
"loss": 0.2235,
"step": 333
},
{
"epoch": 0.6805909322465614,
"grad_norm": 1.115402340888977,
"learning_rate": 4.29811320754717e-06,
"loss": 0.2241,
"step": 334
},
{
"epoch": 0.6826286296484972,
"grad_norm": 1.1034786701202393,
"learning_rate": 4.294339622641509e-06,
"loss": 0.2397,
"step": 335
},
{
"epoch": 0.6846663270504331,
"grad_norm": 1.1569246053695679,
"learning_rate": 4.290566037735849e-06,
"loss": 0.231,
"step": 336
},
{
"epoch": 0.6867040244523688,
"grad_norm": 1.0261273384094238,
"learning_rate": 4.286792452830189e-06,
"loss": 0.2381,
"step": 337
},
{
"epoch": 0.6887417218543046,
"grad_norm": 1.1715890169143677,
"learning_rate": 4.283018867924529e-06,
"loss": 0.2271,
"step": 338
},
{
"epoch": 0.6907794192562404,
"grad_norm": 1.1164259910583496,
"learning_rate": 4.279245283018868e-06,
"loss": 0.2145,
"step": 339
},
{
"epoch": 0.6928171166581762,
"grad_norm": 1.1052844524383545,
"learning_rate": 4.275471698113208e-06,
"loss": 0.2303,
"step": 340
},
{
"epoch": 0.6948548140601121,
"grad_norm": 1.193002700805664,
"learning_rate": 4.271698113207547e-06,
"loss": 0.2329,
"step": 341
},
{
"epoch": 0.6968925114620479,
"grad_norm": 1.141808032989502,
"learning_rate": 4.267924528301887e-06,
"loss": 0.2293,
"step": 342
},
{
"epoch": 0.6989302088639837,
"grad_norm": 1.0740857124328613,
"learning_rate": 4.264150943396227e-06,
"loss": 0.2336,
"step": 343
},
{
"epoch": 0.7009679062659195,
"grad_norm": 1.1825228929519653,
"learning_rate": 4.260377358490567e-06,
"loss": 0.2367,
"step": 344
},
{
"epoch": 0.7030056036678554,
"grad_norm": 1.0624991655349731,
"learning_rate": 4.256603773584906e-06,
"loss": 0.2246,
"step": 345
},
{
"epoch": 0.7050433010697912,
"grad_norm": 1.292654037475586,
"learning_rate": 4.252830188679246e-06,
"loss": 0.2479,
"step": 346
},
{
"epoch": 0.7070809984717269,
"grad_norm": 1.0635449886322021,
"learning_rate": 4.249056603773585e-06,
"loss": 0.2285,
"step": 347
},
{
"epoch": 0.7091186958736627,
"grad_norm": 1.0410432815551758,
"learning_rate": 4.245283018867925e-06,
"loss": 0.203,
"step": 348
},
{
"epoch": 0.7111563932755985,
"grad_norm": 1.154789924621582,
"learning_rate": 4.241509433962264e-06,
"loss": 0.2424,
"step": 349
},
{
"epoch": 0.7131940906775344,
"grad_norm": 1.1573512554168701,
"learning_rate": 4.237735849056604e-06,
"loss": 0.2163,
"step": 350
},
{
"epoch": 0.7152317880794702,
"grad_norm": 1.0690231323242188,
"learning_rate": 4.233962264150944e-06,
"loss": 0.2204,
"step": 351
},
{
"epoch": 0.717269485481406,
"grad_norm": 1.1083498001098633,
"learning_rate": 4.230188679245284e-06,
"loss": 0.2258,
"step": 352
},
{
"epoch": 0.7193071828833418,
"grad_norm": 1.260735273361206,
"learning_rate": 4.226415094339623e-06,
"loss": 0.236,
"step": 353
},
{
"epoch": 0.7213448802852777,
"grad_norm": 1.0777976512908936,
"learning_rate": 4.222641509433963e-06,
"loss": 0.217,
"step": 354
},
{
"epoch": 0.7233825776872135,
"grad_norm": 1.0879008769989014,
"learning_rate": 4.218867924528302e-06,
"loss": 0.2163,
"step": 355
},
{
"epoch": 0.7254202750891493,
"grad_norm": 1.1055690050125122,
"learning_rate": 4.215094339622642e-06,
"loss": 0.2244,
"step": 356
},
{
"epoch": 0.7274579724910851,
"grad_norm": 1.1160818338394165,
"learning_rate": 4.2113207547169815e-06,
"loss": 0.2081,
"step": 357
},
{
"epoch": 0.7294956698930208,
"grad_norm": 1.238552212715149,
"learning_rate": 4.2075471698113215e-06,
"loss": 0.2242,
"step": 358
},
{
"epoch": 0.7315333672949567,
"grad_norm": 1.0889108180999756,
"learning_rate": 4.2037735849056605e-06,
"loss": 0.2276,
"step": 359
},
{
"epoch": 0.7335710646968925,
"grad_norm": 1.314106822013855,
"learning_rate": 4.2000000000000004e-06,
"loss": 0.2423,
"step": 360
},
{
"epoch": 0.7356087620988283,
"grad_norm": 1.304366111755371,
"learning_rate": 4.1962264150943395e-06,
"loss": 0.2556,
"step": 361
},
{
"epoch": 0.7376464595007641,
"grad_norm": 1.227425217628479,
"learning_rate": 4.1924528301886794e-06,
"loss": 0.2275,
"step": 362
},
{
"epoch": 0.7396841569027,
"grad_norm": 1.1975058317184448,
"learning_rate": 4.188679245283019e-06,
"loss": 0.2376,
"step": 363
},
{
"epoch": 0.7417218543046358,
"grad_norm": 1.1609851121902466,
"learning_rate": 4.184905660377358e-06,
"loss": 0.2296,
"step": 364
},
{
"epoch": 0.7437595517065716,
"grad_norm": 1.1305787563323975,
"learning_rate": 4.181132075471698e-06,
"loss": 0.231,
"step": 365
},
{
"epoch": 0.7457972491085074,
"grad_norm": 1.245123267173767,
"learning_rate": 4.177358490566038e-06,
"loss": 0.2438,
"step": 366
},
{
"epoch": 0.7478349465104432,
"grad_norm": 1.2077217102050781,
"learning_rate": 4.173584905660378e-06,
"loss": 0.2331,
"step": 367
},
{
"epoch": 0.7498726439123791,
"grad_norm": 1.2838149070739746,
"learning_rate": 4.169811320754717e-06,
"loss": 0.2205,
"step": 368
},
{
"epoch": 0.7519103413143148,
"grad_norm": 1.2761950492858887,
"learning_rate": 4.166037735849056e-06,
"loss": 0.2339,
"step": 369
},
{
"epoch": 0.7539480387162506,
"grad_norm": 1.2258546352386475,
"learning_rate": 4.162264150943396e-06,
"loss": 0.2494,
"step": 370
},
{
"epoch": 0.7559857361181864,
"grad_norm": 1.0878491401672363,
"learning_rate": 4.158490566037736e-06,
"loss": 0.2337,
"step": 371
},
{
"epoch": 0.7580234335201222,
"grad_norm": 1.4389631748199463,
"learning_rate": 4.154716981132076e-06,
"loss": 0.2409,
"step": 372
},
{
"epoch": 0.7600611309220581,
"grad_norm": 1.0960638523101807,
"learning_rate": 4.150943396226416e-06,
"loss": 0.2239,
"step": 373
},
{
"epoch": 0.7620988283239939,
"grad_norm": 1.293862223625183,
"learning_rate": 4.147169811320755e-06,
"loss": 0.2438,
"step": 374
},
{
"epoch": 0.7641365257259297,
"grad_norm": 1.177188754081726,
"learning_rate": 4.143396226415095e-06,
"loss": 0.2241,
"step": 375
},
{
"epoch": 0.7661742231278655,
"grad_norm": 1.2292778491973877,
"learning_rate": 4.139622641509434e-06,
"loss": 0.2386,
"step": 376
},
{
"epoch": 0.7682119205298014,
"grad_norm": 1.1312750577926636,
"learning_rate": 4.135849056603774e-06,
"loss": 0.2331,
"step": 377
},
{
"epoch": 0.7702496179317372,
"grad_norm": 1.0975465774536133,
"learning_rate": 4.132075471698114e-06,
"loss": 0.2213,
"step": 378
},
{
"epoch": 0.7722873153336729,
"grad_norm": 1.2238826751708984,
"learning_rate": 4.128301886792453e-06,
"loss": 0.2338,
"step": 379
},
{
"epoch": 0.7743250127356087,
"grad_norm": 1.3611332178115845,
"learning_rate": 4.124528301886793e-06,
"loss": 0.2454,
"step": 380
},
{
"epoch": 0.7763627101375445,
"grad_norm": 1.3693833351135254,
"learning_rate": 4.120754716981133e-06,
"loss": 0.2434,
"step": 381
},
{
"epoch": 0.7784004075394804,
"grad_norm": 1.2046077251434326,
"learning_rate": 4.116981132075472e-06,
"loss": 0.224,
"step": 382
},
{
"epoch": 0.7804381049414162,
"grad_norm": 1.2061010599136353,
"learning_rate": 4.113207547169812e-06,
"loss": 0.2264,
"step": 383
},
{
"epoch": 0.782475802343352,
"grad_norm": 1.0464826822280884,
"learning_rate": 4.109433962264151e-06,
"loss": 0.2111,
"step": 384
},
{
"epoch": 0.7845134997452878,
"grad_norm": 0.9789960980415344,
"learning_rate": 4.105660377358491e-06,
"loss": 0.2082,
"step": 385
},
{
"epoch": 0.7865511971472237,
"grad_norm": 1.1676138639450073,
"learning_rate": 4.101886792452831e-06,
"loss": 0.2283,
"step": 386
},
{
"epoch": 0.7885888945491595,
"grad_norm": 1.179202914237976,
"learning_rate": 4.0981132075471705e-06,
"loss": 0.2329,
"step": 387
},
{
"epoch": 0.7906265919510953,
"grad_norm": 1.2767287492752075,
"learning_rate": 4.09433962264151e-06,
"loss": 0.2378,
"step": 388
},
{
"epoch": 0.7926642893530311,
"grad_norm": 1.1678310632705688,
"learning_rate": 4.0905660377358495e-06,
"loss": 0.2232,
"step": 389
},
{
"epoch": 0.7947019867549668,
"grad_norm": 1.2610273361206055,
"learning_rate": 4.0867924528301886e-06,
"loss": 0.2384,
"step": 390
},
{
"epoch": 0.7967396841569027,
"grad_norm": 1.3496994972229004,
"learning_rate": 4.0830188679245285e-06,
"loss": 0.2346,
"step": 391
},
{
"epoch": 0.7987773815588385,
"grad_norm": 1.163509488105774,
"learning_rate": 4.079245283018868e-06,
"loss": 0.2234,
"step": 392
},
{
"epoch": 0.8008150789607743,
"grad_norm": 1.1540744304656982,
"learning_rate": 4.075471698113208e-06,
"loss": 0.2164,
"step": 393
},
{
"epoch": 0.8028527763627101,
"grad_norm": 1.158379316329956,
"learning_rate": 4.071698113207547e-06,
"loss": 0.2323,
"step": 394
},
{
"epoch": 0.804890473764646,
"grad_norm": 1.1848655939102173,
"learning_rate": 4.067924528301887e-06,
"loss": 0.2448,
"step": 395
},
{
"epoch": 0.8069281711665818,
"grad_norm": 1.239961862564087,
"learning_rate": 4.064150943396226e-06,
"loss": 0.2343,
"step": 396
},
{
"epoch": 0.8089658685685176,
"grad_norm": 1.0600473880767822,
"learning_rate": 4.060377358490566e-06,
"loss": 0.2333,
"step": 397
},
{
"epoch": 0.8110035659704534,
"grad_norm": 1.2741254568099976,
"learning_rate": 4.056603773584906e-06,
"loss": 0.237,
"step": 398
},
{
"epoch": 0.8130412633723892,
"grad_norm": 1.182904839515686,
"learning_rate": 4.052830188679245e-06,
"loss": 0.2217,
"step": 399
},
{
"epoch": 0.8150789607743251,
"grad_norm": 1.1751116514205933,
"learning_rate": 4.049056603773585e-06,
"loss": 0.2354,
"step": 400
},
{
"epoch": 0.8171166581762608,
"grad_norm": 1.134203553199768,
"learning_rate": 4.045283018867925e-06,
"loss": 0.2297,
"step": 401
},
{
"epoch": 0.8191543555781966,
"grad_norm": 1.1873515844345093,
"learning_rate": 4.041509433962265e-06,
"loss": 0.2476,
"step": 402
},
{
"epoch": 0.8211920529801324,
"grad_norm": 1.1874173879623413,
"learning_rate": 4.037735849056604e-06,
"loss": 0.2232,
"step": 403
},
{
"epoch": 0.8232297503820682,
"grad_norm": 1.119139552116394,
"learning_rate": 4.033962264150943e-06,
"loss": 0.2348,
"step": 404
},
{
"epoch": 0.8252674477840041,
"grad_norm": 1.1560324430465698,
"learning_rate": 4.030188679245283e-06,
"loss": 0.2337,
"step": 405
},
{
"epoch": 0.8273051451859399,
"grad_norm": 1.1288225650787354,
"learning_rate": 4.026415094339623e-06,
"loss": 0.2319,
"step": 406
},
{
"epoch": 0.8293428425878757,
"grad_norm": 1.2800090312957764,
"learning_rate": 4.022641509433963e-06,
"loss": 0.2237,
"step": 407
},
{
"epoch": 0.8313805399898115,
"grad_norm": 1.2394243478775024,
"learning_rate": 4.018867924528303e-06,
"loss": 0.2358,
"step": 408
},
{
"epoch": 0.8334182373917474,
"grad_norm": 1.231703758239746,
"learning_rate": 4.015094339622642e-06,
"loss": 0.2275,
"step": 409
},
{
"epoch": 0.8354559347936832,
"grad_norm": 1.0887949466705322,
"learning_rate": 4.011320754716982e-06,
"loss": 0.2335,
"step": 410
},
{
"epoch": 0.8374936321956189,
"grad_norm": 1.2228186130523682,
"learning_rate": 4.007547169811321e-06,
"loss": 0.2307,
"step": 411
},
{
"epoch": 0.8395313295975547,
"grad_norm": 1.0364912748336792,
"learning_rate": 4.003773584905661e-06,
"loss": 0.2117,
"step": 412
},
{
"epoch": 0.8415690269994905,
"grad_norm": 1.0746346712112427,
"learning_rate": 4.000000000000001e-06,
"loss": 0.2181,
"step": 413
},
{
"epoch": 0.8436067244014264,
"grad_norm": 1.0695878267288208,
"learning_rate": 3.99622641509434e-06,
"loss": 0.2497,
"step": 414
},
{
"epoch": 0.8456444218033622,
"grad_norm": 1.2379292249679565,
"learning_rate": 3.99245283018868e-06,
"loss": 0.2236,
"step": 415
},
{
"epoch": 0.847682119205298,
"grad_norm": 1.0842210054397583,
"learning_rate": 3.9886792452830196e-06,
"loss": 0.2268,
"step": 416
},
{
"epoch": 0.8497198166072338,
"grad_norm": 1.2367124557495117,
"learning_rate": 3.984905660377359e-06,
"loss": 0.2366,
"step": 417
},
{
"epoch": 0.8517575140091697,
"grad_norm": 1.2747502326965332,
"learning_rate": 3.9811320754716985e-06,
"loss": 0.2371,
"step": 418
},
{
"epoch": 0.8537952114111055,
"grad_norm": 1.1272820234298706,
"learning_rate": 3.977358490566038e-06,
"loss": 0.2369,
"step": 419
},
{
"epoch": 0.8558329088130413,
"grad_norm": 1.0960078239440918,
"learning_rate": 3.9735849056603775e-06,
"loss": 0.2383,
"step": 420
},
{
"epoch": 0.8578706062149771,
"grad_norm": 1.1670606136322021,
"learning_rate": 3.9698113207547174e-06,
"loss": 0.2511,
"step": 421
},
{
"epoch": 0.8599083036169128,
"grad_norm": 1.0942180156707764,
"learning_rate": 3.966037735849057e-06,
"loss": 0.2319,
"step": 422
},
{
"epoch": 0.8619460010188487,
"grad_norm": 1.1233775615692139,
"learning_rate": 3.962264150943396e-06,
"loss": 0.2144,
"step": 423
},
{
"epoch": 0.8639836984207845,
"grad_norm": 1.2059624195098877,
"learning_rate": 3.958490566037736e-06,
"loss": 0.2212,
"step": 424
},
{
"epoch": 0.8660213958227203,
"grad_norm": 1.1963043212890625,
"learning_rate": 3.954716981132075e-06,
"loss": 0.2378,
"step": 425
},
{
"epoch": 0.8680590932246561,
"grad_norm": 1.2415270805358887,
"learning_rate": 3.950943396226415e-06,
"loss": 0.2276,
"step": 426
},
{
"epoch": 0.870096790626592,
"grad_norm": 1.3280036449432373,
"learning_rate": 3.947169811320755e-06,
"loss": 0.2395,
"step": 427
},
{
"epoch": 0.8721344880285278,
"grad_norm": 1.2570695877075195,
"learning_rate": 3.943396226415095e-06,
"loss": 0.2474,
"step": 428
},
{
"epoch": 0.8741721854304636,
"grad_norm": 1.1252264976501465,
"learning_rate": 3.939622641509434e-06,
"loss": 0.2265,
"step": 429
},
{
"epoch": 0.8762098828323994,
"grad_norm": 1.0487228631973267,
"learning_rate": 3.935849056603774e-06,
"loss": 0.2224,
"step": 430
},
{
"epoch": 0.8782475802343352,
"grad_norm": 1.0646063089370728,
"learning_rate": 3.932075471698113e-06,
"loss": 0.2232,
"step": 431
},
{
"epoch": 0.8802852776362711,
"grad_norm": 1.1609469652175903,
"learning_rate": 3.928301886792453e-06,
"loss": 0.2347,
"step": 432
},
{
"epoch": 0.8823229750382068,
"grad_norm": 1.0545512437820435,
"learning_rate": 3.924528301886793e-06,
"loss": 0.2251,
"step": 433
},
{
"epoch": 0.8843606724401426,
"grad_norm": 1.1264142990112305,
"learning_rate": 3.920754716981132e-06,
"loss": 0.2459,
"step": 434
},
{
"epoch": 0.8863983698420784,
"grad_norm": 1.1396156549453735,
"learning_rate": 3.916981132075472e-06,
"loss": 0.2385,
"step": 435
},
{
"epoch": 0.8884360672440142,
"grad_norm": 1.17756187915802,
"learning_rate": 3.913207547169812e-06,
"loss": 0.2306,
"step": 436
},
{
"epoch": 0.8904737646459501,
"grad_norm": 1.0548409223556519,
"learning_rate": 3.909433962264151e-06,
"loss": 0.2192,
"step": 437
},
{
"epoch": 0.8925114620478859,
"grad_norm": 1.161879062652588,
"learning_rate": 3.905660377358491e-06,
"loss": 0.2264,
"step": 438
},
{
"epoch": 0.8945491594498217,
"grad_norm": 1.1480745077133179,
"learning_rate": 3.90188679245283e-06,
"loss": 0.2389,
"step": 439
},
{
"epoch": 0.8965868568517575,
"grad_norm": 1.0667020082473755,
"learning_rate": 3.89811320754717e-06,
"loss": 0.2312,
"step": 440
},
{
"epoch": 0.8986245542536934,
"grad_norm": 1.2451261281967163,
"learning_rate": 3.89433962264151e-06,
"loss": 0.241,
"step": 441
},
{
"epoch": 0.9006622516556292,
"grad_norm": 1.2452954053878784,
"learning_rate": 3.89056603773585e-06,
"loss": 0.2444,
"step": 442
},
{
"epoch": 0.9026999490575649,
"grad_norm": 1.134698510169983,
"learning_rate": 3.88679245283019e-06,
"loss": 0.2132,
"step": 443
},
{
"epoch": 0.9047376464595007,
"grad_norm": 1.269184947013855,
"learning_rate": 3.883018867924529e-06,
"loss": 0.2445,
"step": 444
},
{
"epoch": 0.9067753438614365,
"grad_norm": 1.2156351804733276,
"learning_rate": 3.879245283018868e-06,
"loss": 0.2469,
"step": 445
},
{
"epoch": 0.9088130412633724,
"grad_norm": 1.1011265516281128,
"learning_rate": 3.875471698113208e-06,
"loss": 0.2307,
"step": 446
},
{
"epoch": 0.9108507386653082,
"grad_norm": 1.08492910861969,
"learning_rate": 3.871698113207548e-06,
"loss": 0.2228,
"step": 447
},
{
"epoch": 0.912888436067244,
"grad_norm": 1.1414035558700562,
"learning_rate": 3.8679245283018875e-06,
"loss": 0.2191,
"step": 448
},
{
"epoch": 0.9149261334691798,
"grad_norm": 1.0980679988861084,
"learning_rate": 3.8641509433962266e-06,
"loss": 0.2323,
"step": 449
},
{
"epoch": 0.9169638308711157,
"grad_norm": 1.1721632480621338,
"learning_rate": 3.8603773584905665e-06,
"loss": 0.2457,
"step": 450
},
{
"epoch": 0.9190015282730515,
"grad_norm": 1.1284496784210205,
"learning_rate": 3.856603773584906e-06,
"loss": 0.2326,
"step": 451
},
{
"epoch": 0.9210392256749873,
"grad_norm": 1.0117298364639282,
"learning_rate": 3.8528301886792455e-06,
"loss": 0.2389,
"step": 452
},
{
"epoch": 0.9230769230769231,
"grad_norm": 1.173325777053833,
"learning_rate": 3.849056603773585e-06,
"loss": 0.2304,
"step": 453
},
{
"epoch": 0.9251146204788588,
"grad_norm": 1.0675781965255737,
"learning_rate": 3.8452830188679245e-06,
"loss": 0.2178,
"step": 454
},
{
"epoch": 0.9271523178807947,
"grad_norm": 1.0862107276916504,
"learning_rate": 3.841509433962264e-06,
"loss": 0.2293,
"step": 455
},
{
"epoch": 0.9291900152827305,
"grad_norm": 1.119224190711975,
"learning_rate": 3.837735849056604e-06,
"loss": 0.2228,
"step": 456
},
{
"epoch": 0.9312277126846663,
"grad_norm": 1.0795427560806274,
"learning_rate": 3.833962264150944e-06,
"loss": 0.2235,
"step": 457
},
{
"epoch": 0.9332654100866021,
"grad_norm": 1.1415457725524902,
"learning_rate": 3.830188679245283e-06,
"loss": 0.2272,
"step": 458
},
{
"epoch": 0.935303107488538,
"grad_norm": 1.1307644844055176,
"learning_rate": 3.826415094339623e-06,
"loss": 0.2186,
"step": 459
},
{
"epoch": 0.9373408048904738,
"grad_norm": 1.1211094856262207,
"learning_rate": 3.822641509433962e-06,
"loss": 0.2233,
"step": 460
},
{
"epoch": 0.9393785022924096,
"grad_norm": 1.1230515241622925,
"learning_rate": 3.818867924528302e-06,
"loss": 0.2318,
"step": 461
},
{
"epoch": 0.9414161996943454,
"grad_norm": 1.2053518295288086,
"learning_rate": 3.815094339622642e-06,
"loss": 0.2225,
"step": 462
},
{
"epoch": 0.9434538970962812,
"grad_norm": 1.1487395763397217,
"learning_rate": 3.8113207547169816e-06,
"loss": 0.23,
"step": 463
},
{
"epoch": 0.9454915944982171,
"grad_norm": 1.03309166431427,
"learning_rate": 3.807547169811321e-06,
"loss": 0.2163,
"step": 464
},
{
"epoch": 0.9475292919001528,
"grad_norm": 1.2096184492111206,
"learning_rate": 3.8037735849056605e-06,
"loss": 0.2312,
"step": 465
},
{
"epoch": 0.9495669893020886,
"grad_norm": 1.5864837169647217,
"learning_rate": 3.8000000000000005e-06,
"loss": 0.227,
"step": 466
},
{
"epoch": 0.9516046867040244,
"grad_norm": 1.1054576635360718,
"learning_rate": 3.79622641509434e-06,
"loss": 0.2303,
"step": 467
},
{
"epoch": 0.9536423841059603,
"grad_norm": 1.0742146968841553,
"learning_rate": 3.79245283018868e-06,
"loss": 0.2282,
"step": 468
},
{
"epoch": 0.9556800815078961,
"grad_norm": 1.048632025718689,
"learning_rate": 3.788679245283019e-06,
"loss": 0.222,
"step": 469
},
{
"epoch": 0.9577177789098319,
"grad_norm": 1.1467828750610352,
"learning_rate": 3.784905660377359e-06,
"loss": 0.2169,
"step": 470
},
{
"epoch": 0.9597554763117677,
"grad_norm": 1.1006637811660767,
"learning_rate": 3.7811320754716983e-06,
"loss": 0.227,
"step": 471
},
{
"epoch": 0.9617931737137035,
"grad_norm": 1.4877111911773682,
"learning_rate": 3.7773584905660383e-06,
"loss": 0.2207,
"step": 472
},
{
"epoch": 0.9638308711156394,
"grad_norm": 1.174248456954956,
"learning_rate": 3.7735849056603777e-06,
"loss": 0.2257,
"step": 473
},
{
"epoch": 0.9658685685175752,
"grad_norm": 1.0940933227539062,
"learning_rate": 3.7698113207547172e-06,
"loss": 0.2265,
"step": 474
},
{
"epoch": 0.9679062659195109,
"grad_norm": 1.0824356079101562,
"learning_rate": 3.7660377358490567e-06,
"loss": 0.2261,
"step": 475
},
{
"epoch": 0.9699439633214467,
"grad_norm": 1.0655136108398438,
"learning_rate": 3.7622641509433966e-06,
"loss": 0.2148,
"step": 476
},
{
"epoch": 0.9719816607233825,
"grad_norm": 1.3165481090545654,
"learning_rate": 3.758490566037736e-06,
"loss": 0.2337,
"step": 477
},
{
"epoch": 0.9740193581253184,
"grad_norm": 1.0988367795944214,
"learning_rate": 3.754716981132076e-06,
"loss": 0.1979,
"step": 478
},
{
"epoch": 0.9760570555272542,
"grad_norm": 1.0447558164596558,
"learning_rate": 3.750943396226415e-06,
"loss": 0.2325,
"step": 479
},
{
"epoch": 0.97809475292919,
"grad_norm": 1.1018916368484497,
"learning_rate": 3.747169811320755e-06,
"loss": 0.2161,
"step": 480
},
{
"epoch": 0.9801324503311258,
"grad_norm": 1.2155579328536987,
"learning_rate": 3.7433962264150945e-06,
"loss": 0.2191,
"step": 481
},
{
"epoch": 0.9821701477330617,
"grad_norm": 0.9788108468055725,
"learning_rate": 3.7396226415094344e-06,
"loss": 0.2282,
"step": 482
},
{
"epoch": 0.9842078451349975,
"grad_norm": 1.0340372323989868,
"learning_rate": 3.7358490566037735e-06,
"loss": 0.2276,
"step": 483
},
{
"epoch": 0.9862455425369333,
"grad_norm": 0.9971087574958801,
"learning_rate": 3.7320754716981134e-06,
"loss": 0.2176,
"step": 484
},
{
"epoch": 0.9882832399388691,
"grad_norm": 1.0751736164093018,
"learning_rate": 3.728301886792453e-06,
"loss": 0.2143,
"step": 485
},
{
"epoch": 0.9903209373408048,
"grad_norm": 1.188984751701355,
"learning_rate": 3.724528301886793e-06,
"loss": 0.2375,
"step": 486
},
{
"epoch": 0.9923586347427407,
"grad_norm": 1.320594072341919,
"learning_rate": 3.7207547169811327e-06,
"loss": 0.223,
"step": 487
},
{
"epoch": 0.9943963321446765,
"grad_norm": 1.1396737098693848,
"learning_rate": 3.716981132075472e-06,
"loss": 0.2413,
"step": 488
},
{
"epoch": 0.9964340295466123,
"grad_norm": 1.0497945547103882,
"learning_rate": 3.7132075471698113e-06,
"loss": 0.2177,
"step": 489
},
{
"epoch": 0.9984717269485481,
"grad_norm": 1.2380748987197876,
"learning_rate": 3.709433962264151e-06,
"loss": 0.2351,
"step": 490
},
{
"epoch": 1.000509424350484,
"grad_norm": 0.9542668461799622,
"learning_rate": 3.705660377358491e-06,
"loss": 0.2136,
"step": 491
},
{
"epoch": 1.0025471217524198,
"grad_norm": 0.9574536681175232,
"learning_rate": 3.7018867924528306e-06,
"loss": 0.1899,
"step": 492
},
{
"epoch": 1.0045848191543556,
"grad_norm": 1.0352755784988403,
"learning_rate": 3.6981132075471697e-06,
"loss": 0.1816,
"step": 493
},
{
"epoch": 1.0066225165562914,
"grad_norm": 1.0826165676116943,
"learning_rate": 3.6943396226415096e-06,
"loss": 0.1858,
"step": 494
},
{
"epoch": 1.0086602139582272,
"grad_norm": 1.2422000169754028,
"learning_rate": 3.6905660377358495e-06,
"loss": 0.1878,
"step": 495
},
{
"epoch": 1.010697911360163,
"grad_norm": 1.0961295366287231,
"learning_rate": 3.686792452830189e-06,
"loss": 0.1721,
"step": 496
},
{
"epoch": 1.012735608762099,
"grad_norm": 1.2105534076690674,
"learning_rate": 3.683018867924529e-06,
"loss": 0.1682,
"step": 497
},
{
"epoch": 1.0147733061640347,
"grad_norm": 1.0163434743881226,
"learning_rate": 3.679245283018868e-06,
"loss": 0.1745,
"step": 498
},
{
"epoch": 1.0168110035659705,
"grad_norm": 1.1357200145721436,
"learning_rate": 3.675471698113208e-06,
"loss": 0.174,
"step": 499
},
{
"epoch": 1.0188487009679064,
"grad_norm": 1.129521369934082,
"learning_rate": 3.6716981132075474e-06,
"loss": 0.1737,
"step": 500
},
{
"epoch": 1.0208863983698422,
"grad_norm": 1.1067070960998535,
"learning_rate": 3.6679245283018873e-06,
"loss": 0.1715,
"step": 501
},
{
"epoch": 1.0229240957717778,
"grad_norm": 1.3292362689971924,
"learning_rate": 3.664150943396227e-06,
"loss": 0.1719,
"step": 502
},
{
"epoch": 1.0249617931737136,
"grad_norm": 1.184263825416565,
"learning_rate": 3.6603773584905663e-06,
"loss": 0.1772,
"step": 503
},
{
"epoch": 1.0269994905756494,
"grad_norm": 1.2224076986312866,
"learning_rate": 3.6566037735849058e-06,
"loss": 0.1799,
"step": 504
},
{
"epoch": 1.0290371879775853,
"grad_norm": 1.2455564737319946,
"learning_rate": 3.6528301886792457e-06,
"loss": 0.1875,
"step": 505
},
{
"epoch": 1.031074885379521,
"grad_norm": 1.037973165512085,
"learning_rate": 3.649056603773585e-06,
"loss": 0.1751,
"step": 506
},
{
"epoch": 1.033112582781457,
"grad_norm": 1.3113584518432617,
"learning_rate": 3.645283018867925e-06,
"loss": 0.1805,
"step": 507
},
{
"epoch": 1.0351502801833927,
"grad_norm": 1.113845705986023,
"learning_rate": 3.641509433962264e-06,
"loss": 0.163,
"step": 508
},
{
"epoch": 1.0371879775853285,
"grad_norm": 1.1282869577407837,
"learning_rate": 3.637735849056604e-06,
"loss": 0.1774,
"step": 509
},
{
"epoch": 1.0392256749872644,
"grad_norm": 0.9915235042572021,
"learning_rate": 3.6339622641509436e-06,
"loss": 0.1676,
"step": 510
},
{
"epoch": 1.0412633723892002,
"grad_norm": 1.1076091527938843,
"learning_rate": 3.6301886792452835e-06,
"loss": 0.1811,
"step": 511
},
{
"epoch": 1.043301069791136,
"grad_norm": 1.4706580638885498,
"learning_rate": 3.626415094339623e-06,
"loss": 0.1749,
"step": 512
},
{
"epoch": 1.0453387671930718,
"grad_norm": 1.0995841026306152,
"learning_rate": 3.6226415094339625e-06,
"loss": 0.1682,
"step": 513
},
{
"epoch": 1.0473764645950077,
"grad_norm": 1.3873177766799927,
"learning_rate": 3.618867924528302e-06,
"loss": 0.1812,
"step": 514
},
{
"epoch": 1.0494141619969435,
"grad_norm": 1.1935499906539917,
"learning_rate": 3.615094339622642e-06,
"loss": 0.1876,
"step": 515
},
{
"epoch": 1.0514518593988793,
"grad_norm": 1.2057229280471802,
"learning_rate": 3.6113207547169814e-06,
"loss": 0.1815,
"step": 516
},
{
"epoch": 1.0534895568008151,
"grad_norm": 1.1333197355270386,
"learning_rate": 3.6075471698113213e-06,
"loss": 0.1828,
"step": 517
},
{
"epoch": 1.055527254202751,
"grad_norm": 1.0647273063659668,
"learning_rate": 3.6037735849056603e-06,
"loss": 0.181,
"step": 518
},
{
"epoch": 1.0575649516046868,
"grad_norm": 1.204564094543457,
"learning_rate": 3.6000000000000003e-06,
"loss": 0.1825,
"step": 519
},
{
"epoch": 1.0596026490066226,
"grad_norm": 1.0661295652389526,
"learning_rate": 3.5962264150943398e-06,
"loss": 0.1864,
"step": 520
},
{
"epoch": 1.0616403464085584,
"grad_norm": 1.0870025157928467,
"learning_rate": 3.5924528301886797e-06,
"loss": 0.1789,
"step": 521
},
{
"epoch": 1.0636780438104942,
"grad_norm": 1.0620194673538208,
"learning_rate": 3.588679245283019e-06,
"loss": 0.1863,
"step": 522
},
{
"epoch": 1.06571574121243,
"grad_norm": 1.1938071250915527,
"learning_rate": 3.5849056603773586e-06,
"loss": 0.1863,
"step": 523
},
{
"epoch": 1.0677534386143657,
"grad_norm": 1.2299485206604004,
"learning_rate": 3.581132075471698e-06,
"loss": 0.1881,
"step": 524
},
{
"epoch": 1.0697911360163015,
"grad_norm": 1.043164610862732,
"learning_rate": 3.577358490566038e-06,
"loss": 0.1631,
"step": 525
},
{
"epoch": 1.0718288334182373,
"grad_norm": 1.200393795967102,
"learning_rate": 3.5735849056603775e-06,
"loss": 0.1871,
"step": 526
},
{
"epoch": 1.0738665308201731,
"grad_norm": 1.1729276180267334,
"learning_rate": 3.5698113207547175e-06,
"loss": 0.1776,
"step": 527
},
{
"epoch": 1.075904228222109,
"grad_norm": 1.3533014059066772,
"learning_rate": 3.5660377358490565e-06,
"loss": 0.1924,
"step": 528
},
{
"epoch": 1.0779419256240448,
"grad_norm": 1.1192210912704468,
"learning_rate": 3.5622641509433964e-06,
"loss": 0.1826,
"step": 529
},
{
"epoch": 1.0799796230259806,
"grad_norm": 1.2234528064727783,
"learning_rate": 3.558490566037736e-06,
"loss": 0.1803,
"step": 530
},
{
"epoch": 1.0820173204279164,
"grad_norm": 1.1349793672561646,
"learning_rate": 3.554716981132076e-06,
"loss": 0.1862,
"step": 531
},
{
"epoch": 1.0840550178298523,
"grad_norm": 1.1058518886566162,
"learning_rate": 3.5509433962264158e-06,
"loss": 0.1722,
"step": 532
},
{
"epoch": 1.086092715231788,
"grad_norm": 1.0707038640975952,
"learning_rate": 3.547169811320755e-06,
"loss": 0.1709,
"step": 533
},
{
"epoch": 1.088130412633724,
"grad_norm": 1.2310295104980469,
"learning_rate": 3.5433962264150943e-06,
"loss": 0.187,
"step": 534
},
{
"epoch": 1.0901681100356597,
"grad_norm": 1.098715901374817,
"learning_rate": 3.5396226415094342e-06,
"loss": 0.1695,
"step": 535
},
{
"epoch": 1.0922058074375955,
"grad_norm": 1.1150951385498047,
"learning_rate": 3.535849056603774e-06,
"loss": 0.1717,
"step": 536
},
{
"epoch": 1.0942435048395314,
"grad_norm": 1.0338242053985596,
"learning_rate": 3.5320754716981136e-06,
"loss": 0.1789,
"step": 537
},
{
"epoch": 1.0962812022414672,
"grad_norm": 1.0984159708023071,
"learning_rate": 3.5283018867924527e-06,
"loss": 0.1767,
"step": 538
},
{
"epoch": 1.098318899643403,
"grad_norm": 1.1370503902435303,
"learning_rate": 3.5245283018867926e-06,
"loss": 0.1863,
"step": 539
},
{
"epoch": 1.1003565970453388,
"grad_norm": 1.1123195886611938,
"learning_rate": 3.5207547169811325e-06,
"loss": 0.1733,
"step": 540
},
{
"epoch": 1.1023942944472747,
"grad_norm": 1.1519520282745361,
"learning_rate": 3.516981132075472e-06,
"loss": 0.1669,
"step": 541
},
{
"epoch": 1.1044319918492105,
"grad_norm": 1.1219109296798706,
"learning_rate": 3.513207547169812e-06,
"loss": 0.1788,
"step": 542
},
{
"epoch": 1.1064696892511463,
"grad_norm": 1.237865686416626,
"learning_rate": 3.509433962264151e-06,
"loss": 0.1776,
"step": 543
},
{
"epoch": 1.108507386653082,
"grad_norm": 1.0959861278533936,
"learning_rate": 3.505660377358491e-06,
"loss": 0.1773,
"step": 544
},
{
"epoch": 1.1105450840550177,
"grad_norm": 1.079746127128601,
"learning_rate": 3.5018867924528304e-06,
"loss": 0.1942,
"step": 545
},
{
"epoch": 1.1125827814569536,
"grad_norm": 1.1233259439468384,
"learning_rate": 3.4981132075471703e-06,
"loss": 0.1707,
"step": 546
},
{
"epoch": 1.1146204788588894,
"grad_norm": 1.2879219055175781,
"learning_rate": 3.49433962264151e-06,
"loss": 0.1746,
"step": 547
},
{
"epoch": 1.1166581762608252,
"grad_norm": 1.1267422437667847,
"learning_rate": 3.4905660377358493e-06,
"loss": 0.1782,
"step": 548
},
{
"epoch": 1.118695873662761,
"grad_norm": 1.397052526473999,
"learning_rate": 3.486792452830189e-06,
"loss": 0.1696,
"step": 549
},
{
"epoch": 1.1207335710646968,
"grad_norm": 1.3258302211761475,
"learning_rate": 3.4830188679245287e-06,
"loss": 0.1668,
"step": 550
},
{
"epoch": 1.1227712684666327,
"grad_norm": 1.225081205368042,
"learning_rate": 3.479245283018868e-06,
"loss": 0.179,
"step": 551
},
{
"epoch": 1.1248089658685685,
"grad_norm": 1.187245488166809,
"learning_rate": 3.475471698113208e-06,
"loss": 0.1739,
"step": 552
},
{
"epoch": 1.1268466632705043,
"grad_norm": 1.2275511026382446,
"learning_rate": 3.471698113207547e-06,
"loss": 0.1892,
"step": 553
},
{
"epoch": 1.1288843606724401,
"grad_norm": 1.1659022569656372,
"learning_rate": 3.467924528301887e-06,
"loss": 0.1801,
"step": 554
},
{
"epoch": 1.130922058074376,
"grad_norm": 1.3677842617034912,
"learning_rate": 3.4641509433962266e-06,
"loss": 0.1787,
"step": 555
},
{
"epoch": 1.1329597554763118,
"grad_norm": 1.2617255449295044,
"learning_rate": 3.4603773584905665e-06,
"loss": 0.1792,
"step": 556
},
{
"epoch": 1.1349974528782476,
"grad_norm": 1.1734035015106201,
"learning_rate": 3.456603773584906e-06,
"loss": 0.1735,
"step": 557
},
{
"epoch": 1.1370351502801834,
"grad_norm": 1.3135229349136353,
"learning_rate": 3.4528301886792455e-06,
"loss": 0.1844,
"step": 558
},
{
"epoch": 1.1390728476821192,
"grad_norm": 1.281538724899292,
"learning_rate": 3.449056603773585e-06,
"loss": 0.1811,
"step": 559
},
{
"epoch": 1.141110545084055,
"grad_norm": 1.1368190050125122,
"learning_rate": 3.445283018867925e-06,
"loss": 0.1633,
"step": 560
},
{
"epoch": 1.143148242485991,
"grad_norm": 1.0890092849731445,
"learning_rate": 3.4415094339622644e-06,
"loss": 0.178,
"step": 561
},
{
"epoch": 1.1451859398879267,
"grad_norm": 1.20881986618042,
"learning_rate": 3.4377358490566043e-06,
"loss": 0.1722,
"step": 562
},
{
"epoch": 1.1472236372898625,
"grad_norm": 1.1593676805496216,
"learning_rate": 3.4339622641509434e-06,
"loss": 0.1829,
"step": 563
},
{
"epoch": 1.1492613346917984,
"grad_norm": 1.090755820274353,
"learning_rate": 3.4301886792452833e-06,
"loss": 0.1787,
"step": 564
},
{
"epoch": 1.1512990320937342,
"grad_norm": 1.2112749814987183,
"learning_rate": 3.4264150943396228e-06,
"loss": 0.1801,
"step": 565
},
{
"epoch": 1.15333672949567,
"grad_norm": 1.119545340538025,
"learning_rate": 3.4226415094339627e-06,
"loss": 0.1821,
"step": 566
},
{
"epoch": 1.1553744268976056,
"grad_norm": 1.1820521354675293,
"learning_rate": 3.4188679245283026e-06,
"loss": 0.189,
"step": 567
},
{
"epoch": 1.1574121242995414,
"grad_norm": 1.2243889570236206,
"learning_rate": 3.4150943396226417e-06,
"loss": 0.1838,
"step": 568
},
{
"epoch": 1.1594498217014773,
"grad_norm": 1.0234663486480713,
"learning_rate": 3.411320754716981e-06,
"loss": 0.1767,
"step": 569
},
{
"epoch": 1.161487519103413,
"grad_norm": 1.209953784942627,
"learning_rate": 3.407547169811321e-06,
"loss": 0.1769,
"step": 570
},
{
"epoch": 1.163525216505349,
"grad_norm": 1.1745116710662842,
"learning_rate": 3.403773584905661e-06,
"loss": 0.1856,
"step": 571
},
{
"epoch": 1.1655629139072847,
"grad_norm": 1.071757435798645,
"learning_rate": 3.4000000000000005e-06,
"loss": 0.1591,
"step": 572
},
{
"epoch": 1.1676006113092205,
"grad_norm": 1.150458574295044,
"learning_rate": 3.3962264150943395e-06,
"loss": 0.1776,
"step": 573
},
{
"epoch": 1.1696383087111564,
"grad_norm": 1.193291187286377,
"learning_rate": 3.3924528301886795e-06,
"loss": 0.175,
"step": 574
},
{
"epoch": 1.1716760061130922,
"grad_norm": 1.2312043905258179,
"learning_rate": 3.3886792452830194e-06,
"loss": 0.1906,
"step": 575
},
{
"epoch": 1.173713703515028,
"grad_norm": 1.1354984045028687,
"learning_rate": 3.384905660377359e-06,
"loss": 0.1753,
"step": 576
},
{
"epoch": 1.1757514009169638,
"grad_norm": 1.3425500392913818,
"learning_rate": 3.3811320754716988e-06,
"loss": 0.1876,
"step": 577
},
{
"epoch": 1.1777890983188997,
"grad_norm": 1.0738446712493896,
"learning_rate": 3.377358490566038e-06,
"loss": 0.1765,
"step": 578
},
{
"epoch": 1.1798267957208355,
"grad_norm": 1.1612354516983032,
"learning_rate": 3.3735849056603773e-06,
"loss": 0.1712,
"step": 579
},
{
"epoch": 1.1818644931227713,
"grad_norm": 1.2308764457702637,
"learning_rate": 3.3698113207547173e-06,
"loss": 0.182,
"step": 580
},
{
"epoch": 1.1839021905247071,
"grad_norm": 1.3299064636230469,
"learning_rate": 3.366037735849057e-06,
"loss": 0.1812,
"step": 581
},
{
"epoch": 1.185939887926643,
"grad_norm": 1.1064029932022095,
"learning_rate": 3.3622641509433967e-06,
"loss": 0.1853,
"step": 582
},
{
"epoch": 1.1879775853285788,
"grad_norm": 1.131239414215088,
"learning_rate": 3.3584905660377357e-06,
"loss": 0.1827,
"step": 583
},
{
"epoch": 1.1900152827305146,
"grad_norm": 1.1805070638656616,
"learning_rate": 3.3547169811320756e-06,
"loss": 0.1937,
"step": 584
},
{
"epoch": 1.1920529801324504,
"grad_norm": 1.2116690874099731,
"learning_rate": 3.3509433962264156e-06,
"loss": 0.197,
"step": 585
},
{
"epoch": 1.194090677534386,
"grad_norm": 1.3518807888031006,
"learning_rate": 3.347169811320755e-06,
"loss": 0.1765,
"step": 586
},
{
"epoch": 1.1961283749363218,
"grad_norm": 1.2591750621795654,
"learning_rate": 3.343396226415095e-06,
"loss": 0.1782,
"step": 587
},
{
"epoch": 1.1981660723382577,
"grad_norm": 1.1681146621704102,
"learning_rate": 3.339622641509434e-06,
"loss": 0.1811,
"step": 588
},
{
"epoch": 1.2002037697401935,
"grad_norm": 1.2340030670166016,
"learning_rate": 3.335849056603774e-06,
"loss": 0.1732,
"step": 589
},
{
"epoch": 1.2022414671421293,
"grad_norm": 1.2480478286743164,
"learning_rate": 3.3320754716981134e-06,
"loss": 0.1747,
"step": 590
},
{
"epoch": 1.2042791645440651,
"grad_norm": 1.2134257555007935,
"learning_rate": 3.3283018867924534e-06,
"loss": 0.1807,
"step": 591
},
{
"epoch": 1.206316861946001,
"grad_norm": 1.050817608833313,
"learning_rate": 3.324528301886793e-06,
"loss": 0.1725,
"step": 592
},
{
"epoch": 1.2083545593479368,
"grad_norm": 1.2634903192520142,
"learning_rate": 3.3207547169811323e-06,
"loss": 0.1883,
"step": 593
},
{
"epoch": 1.2103922567498726,
"grad_norm": 1.2350244522094727,
"learning_rate": 3.316981132075472e-06,
"loss": 0.1872,
"step": 594
},
{
"epoch": 1.2124299541518084,
"grad_norm": 1.232961893081665,
"learning_rate": 3.3132075471698117e-06,
"loss": 0.1654,
"step": 595
},
{
"epoch": 1.2144676515537443,
"grad_norm": 1.163649320602417,
"learning_rate": 3.3094339622641512e-06,
"loss": 0.1933,
"step": 596
},
{
"epoch": 1.21650534895568,
"grad_norm": 1.21866774559021,
"learning_rate": 3.305660377358491e-06,
"loss": 0.1818,
"step": 597
},
{
"epoch": 1.218543046357616,
"grad_norm": 1.1113258600234985,
"learning_rate": 3.30188679245283e-06,
"loss": 0.1755,
"step": 598
},
{
"epoch": 1.2205807437595517,
"grad_norm": 1.1248152256011963,
"learning_rate": 3.29811320754717e-06,
"loss": 0.1756,
"step": 599
},
{
"epoch": 1.2226184411614875,
"grad_norm": 1.1161712408065796,
"learning_rate": 3.2943396226415096e-06,
"loss": 0.1967,
"step": 600
},
{
"epoch": 1.2246561385634234,
"grad_norm": 1.1488161087036133,
"learning_rate": 3.2905660377358495e-06,
"loss": 0.1791,
"step": 601
},
{
"epoch": 1.2266938359653592,
"grad_norm": 1.2753115892410278,
"learning_rate": 3.286792452830189e-06,
"loss": 0.1739,
"step": 602
},
{
"epoch": 1.228731533367295,
"grad_norm": 1.1130990982055664,
"learning_rate": 3.2830188679245285e-06,
"loss": 0.1682,
"step": 603
},
{
"epoch": 1.2307692307692308,
"grad_norm": 1.1455460786819458,
"learning_rate": 3.279245283018868e-06,
"loss": 0.1854,
"step": 604
},
{
"epoch": 1.2328069281711667,
"grad_norm": 1.1896706819534302,
"learning_rate": 3.275471698113208e-06,
"loss": 0.1901,
"step": 605
},
{
"epoch": 1.2348446255731025,
"grad_norm": 1.132242202758789,
"learning_rate": 3.2716981132075474e-06,
"loss": 0.1705,
"step": 606
},
{
"epoch": 1.2368823229750383,
"grad_norm": 1.1296707391738892,
"learning_rate": 3.2679245283018873e-06,
"loss": 0.1875,
"step": 607
},
{
"epoch": 1.2389200203769741,
"grad_norm": 1.2837047576904297,
"learning_rate": 3.2641509433962264e-06,
"loss": 0.1862,
"step": 608
},
{
"epoch": 1.24095771777891,
"grad_norm": 1.2516281604766846,
"learning_rate": 3.2603773584905663e-06,
"loss": 0.1763,
"step": 609
},
{
"epoch": 1.2429954151808456,
"grad_norm": 1.2051138877868652,
"learning_rate": 3.256603773584906e-06,
"loss": 0.1681,
"step": 610
},
{
"epoch": 1.2450331125827814,
"grad_norm": 1.1206097602844238,
"learning_rate": 3.2528301886792457e-06,
"loss": 0.1895,
"step": 611
},
{
"epoch": 1.2470708099847172,
"grad_norm": 1.085570216178894,
"learning_rate": 3.2490566037735848e-06,
"loss": 0.1674,
"step": 612
},
{
"epoch": 1.249108507386653,
"grad_norm": 1.1711559295654297,
"learning_rate": 3.2452830188679247e-06,
"loss": 0.1847,
"step": 613
},
{
"epoch": 1.2511462047885888,
"grad_norm": 1.2569772005081177,
"learning_rate": 3.241509433962264e-06,
"loss": 0.185,
"step": 614
},
{
"epoch": 1.2531839021905247,
"grad_norm": 1.265191912651062,
"learning_rate": 3.237735849056604e-06,
"loss": 0.188,
"step": 615
},
{
"epoch": 1.2552215995924605,
"grad_norm": 1.2143467664718628,
"learning_rate": 3.233962264150944e-06,
"loss": 0.189,
"step": 616
},
{
"epoch": 1.2572592969943963,
"grad_norm": 1.3829542398452759,
"learning_rate": 3.230188679245283e-06,
"loss": 0.174,
"step": 617
},
{
"epoch": 1.2592969943963321,
"grad_norm": 1.2590124607086182,
"learning_rate": 3.2264150943396226e-06,
"loss": 0.1921,
"step": 618
},
{
"epoch": 1.261334691798268,
"grad_norm": 1.125143051147461,
"learning_rate": 3.2226415094339625e-06,
"loss": 0.1788,
"step": 619
},
{
"epoch": 1.2633723892002038,
"grad_norm": 1.136713981628418,
"learning_rate": 3.2188679245283024e-06,
"loss": 0.1665,
"step": 620
},
{
"epoch": 1.2654100866021396,
"grad_norm": 1.1080840826034546,
"learning_rate": 3.215094339622642e-06,
"loss": 0.1785,
"step": 621
},
{
"epoch": 1.2674477840040754,
"grad_norm": 1.0990139245986938,
"learning_rate": 3.211320754716981e-06,
"loss": 0.1759,
"step": 622
},
{
"epoch": 1.2694854814060113,
"grad_norm": 1.2469940185546875,
"learning_rate": 3.207547169811321e-06,
"loss": 0.1738,
"step": 623
},
{
"epoch": 1.271523178807947,
"grad_norm": 1.163061261177063,
"learning_rate": 3.2037735849056608e-06,
"loss": 0.1881,
"step": 624
},
{
"epoch": 1.273560876209883,
"grad_norm": 1.1554782390594482,
"learning_rate": 3.2000000000000003e-06,
"loss": 0.1728,
"step": 625
},
{
"epoch": 1.2755985736118187,
"grad_norm": 1.1074477434158325,
"learning_rate": 3.19622641509434e-06,
"loss": 0.1762,
"step": 626
},
{
"epoch": 1.2776362710137543,
"grad_norm": 1.1363695859909058,
"learning_rate": 3.1924528301886793e-06,
"loss": 0.172,
"step": 627
},
{
"epoch": 1.2796739684156901,
"grad_norm": 1.0740599632263184,
"learning_rate": 3.188679245283019e-06,
"loss": 0.1634,
"step": 628
},
{
"epoch": 1.281711665817626,
"grad_norm": 1.0671052932739258,
"learning_rate": 3.1849056603773587e-06,
"loss": 0.1749,
"step": 629
},
{
"epoch": 1.2837493632195618,
"grad_norm": 1.1366360187530518,
"learning_rate": 3.1811320754716986e-06,
"loss": 0.1685,
"step": 630
},
{
"epoch": 1.2857870606214976,
"grad_norm": 1.250622272491455,
"learning_rate": 3.177358490566038e-06,
"loss": 0.1758,
"step": 631
},
{
"epoch": 1.2878247580234334,
"grad_norm": 1.145407795906067,
"learning_rate": 3.1735849056603776e-06,
"loss": 0.1881,
"step": 632
},
{
"epoch": 1.2898624554253693,
"grad_norm": 1.1561169624328613,
"learning_rate": 3.169811320754717e-06,
"loss": 0.1846,
"step": 633
},
{
"epoch": 1.291900152827305,
"grad_norm": 1.1274852752685547,
"learning_rate": 3.166037735849057e-06,
"loss": 0.1765,
"step": 634
},
{
"epoch": 1.293937850229241,
"grad_norm": 1.2915289402008057,
"learning_rate": 3.1622641509433965e-06,
"loss": 0.1767,
"step": 635
},
{
"epoch": 1.2959755476311767,
"grad_norm": 1.1345237493515015,
"learning_rate": 3.1584905660377364e-06,
"loss": 0.1689,
"step": 636
},
{
"epoch": 1.2980132450331126,
"grad_norm": 1.2380014657974243,
"learning_rate": 3.1547169811320754e-06,
"loss": 0.19,
"step": 637
},
{
"epoch": 1.3000509424350484,
"grad_norm": 1.1787712574005127,
"learning_rate": 3.1509433962264154e-06,
"loss": 0.187,
"step": 638
},
{
"epoch": 1.3020886398369842,
"grad_norm": 1.172777771949768,
"learning_rate": 3.147169811320755e-06,
"loss": 0.1951,
"step": 639
},
{
"epoch": 1.30412633723892,
"grad_norm": 1.1491492986679077,
"learning_rate": 3.1433962264150948e-06,
"loss": 0.1651,
"step": 640
},
{
"epoch": 1.3061640346408558,
"grad_norm": 1.1255732774734497,
"learning_rate": 3.1396226415094343e-06,
"loss": 0.1838,
"step": 641
},
{
"epoch": 1.3082017320427917,
"grad_norm": 1.2315205335617065,
"learning_rate": 3.1358490566037737e-06,
"loss": 0.1785,
"step": 642
},
{
"epoch": 1.3102394294447275,
"grad_norm": 1.1849606037139893,
"learning_rate": 3.1320754716981132e-06,
"loss": 0.177,
"step": 643
},
{
"epoch": 1.3122771268466633,
"grad_norm": 1.1372692584991455,
"learning_rate": 3.128301886792453e-06,
"loss": 0.1747,
"step": 644
},
{
"epoch": 1.3143148242485991,
"grad_norm": 1.2609679698944092,
"learning_rate": 3.1245283018867926e-06,
"loss": 0.1836,
"step": 645
},
{
"epoch": 1.316352521650535,
"grad_norm": 1.179504156112671,
"learning_rate": 3.1207547169811326e-06,
"loss": 0.1771,
"step": 646
},
{
"epoch": 1.3183902190524708,
"grad_norm": 1.2097948789596558,
"learning_rate": 3.1169811320754716e-06,
"loss": 0.1873,
"step": 647
},
{
"epoch": 1.3204279164544066,
"grad_norm": 1.1823457479476929,
"learning_rate": 3.1132075471698115e-06,
"loss": 0.1893,
"step": 648
},
{
"epoch": 1.3224656138563424,
"grad_norm": 1.1036756038665771,
"learning_rate": 3.109433962264151e-06,
"loss": 0.1651,
"step": 649
},
{
"epoch": 1.3245033112582782,
"grad_norm": 1.1787657737731934,
"learning_rate": 3.105660377358491e-06,
"loss": 0.1675,
"step": 650
},
{
"epoch": 1.326541008660214,
"grad_norm": 1.2578370571136475,
"learning_rate": 3.1018867924528304e-06,
"loss": 0.1854,
"step": 651
},
{
"epoch": 1.32857870606215,
"grad_norm": 1.136648178100586,
"learning_rate": 3.09811320754717e-06,
"loss": 0.1744,
"step": 652
},
{
"epoch": 1.3306164034640857,
"grad_norm": 1.1764239072799683,
"learning_rate": 3.0943396226415094e-06,
"loss": 0.1721,
"step": 653
},
{
"epoch": 1.3326541008660213,
"grad_norm": 1.0723998546600342,
"learning_rate": 3.0905660377358493e-06,
"loss": 0.1719,
"step": 654
},
{
"epoch": 1.3346917982679571,
"grad_norm": 1.1238343715667725,
"learning_rate": 3.086792452830189e-06,
"loss": 0.1752,
"step": 655
},
{
"epoch": 1.336729495669893,
"grad_norm": 1.1110343933105469,
"learning_rate": 3.0830188679245287e-06,
"loss": 0.1806,
"step": 656
},
{
"epoch": 1.3387671930718288,
"grad_norm": 1.1625982522964478,
"learning_rate": 3.079245283018868e-06,
"loss": 0.1919,
"step": 657
},
{
"epoch": 1.3408048904737646,
"grad_norm": 1.2139103412628174,
"learning_rate": 3.0754716981132077e-06,
"loss": 0.1807,
"step": 658
},
{
"epoch": 1.3428425878757004,
"grad_norm": 1.2624015808105469,
"learning_rate": 3.071698113207547e-06,
"loss": 0.1774,
"step": 659
},
{
"epoch": 1.3448802852776363,
"grad_norm": 1.1833164691925049,
"learning_rate": 3.067924528301887e-06,
"loss": 0.1765,
"step": 660
},
{
"epoch": 1.346917982679572,
"grad_norm": 1.3421837091445923,
"learning_rate": 3.064150943396227e-06,
"loss": 0.1861,
"step": 661
},
{
"epoch": 1.348955680081508,
"grad_norm": 1.1380257606506348,
"learning_rate": 3.060377358490566e-06,
"loss": 0.1642,
"step": 662
},
{
"epoch": 1.3509933774834437,
"grad_norm": 1.2193323373794556,
"learning_rate": 3.0566037735849056e-06,
"loss": 0.1804,
"step": 663
},
{
"epoch": 1.3530310748853795,
"grad_norm": 1.0917553901672363,
"learning_rate": 3.0528301886792455e-06,
"loss": 0.1761,
"step": 664
},
{
"epoch": 1.3550687722873154,
"grad_norm": 1.252640724182129,
"learning_rate": 3.0490566037735854e-06,
"loss": 0.1892,
"step": 665
},
{
"epoch": 1.3571064696892512,
"grad_norm": 1.2436408996582031,
"learning_rate": 3.045283018867925e-06,
"loss": 0.1865,
"step": 666
},
{
"epoch": 1.359144167091187,
"grad_norm": 1.0737476348876953,
"learning_rate": 3.041509433962264e-06,
"loss": 0.173,
"step": 667
},
{
"epoch": 1.3611818644931228,
"grad_norm": 1.3767677545547485,
"learning_rate": 3.037735849056604e-06,
"loss": 0.1855,
"step": 668
},
{
"epoch": 1.3632195618950587,
"grad_norm": 1.1147671937942505,
"learning_rate": 3.033962264150944e-06,
"loss": 0.175,
"step": 669
},
{
"epoch": 1.3652572592969943,
"grad_norm": 1.2812708616256714,
"learning_rate": 3.0301886792452833e-06,
"loss": 0.1844,
"step": 670
},
{
"epoch": 1.36729495669893,
"grad_norm": 1.028883695602417,
"learning_rate": 3.0264150943396232e-06,
"loss": 0.1641,
"step": 671
},
{
"epoch": 1.369332654100866,
"grad_norm": 1.2508153915405273,
"learning_rate": 3.0226415094339623e-06,
"loss": 0.1884,
"step": 672
},
{
"epoch": 1.3713703515028017,
"grad_norm": 1.2635626792907715,
"learning_rate": 3.018867924528302e-06,
"loss": 0.1835,
"step": 673
},
{
"epoch": 1.3734080489047376,
"grad_norm": 1.1258081197738647,
"learning_rate": 3.0150943396226417e-06,
"loss": 0.1694,
"step": 674
},
{
"epoch": 1.3754457463066734,
"grad_norm": 1.1584776639938354,
"learning_rate": 3.0113207547169816e-06,
"loss": 0.1928,
"step": 675
},
{
"epoch": 1.3774834437086092,
"grad_norm": 1.1394814252853394,
"learning_rate": 3.007547169811321e-06,
"loss": 0.1698,
"step": 676
},
{
"epoch": 1.379521141110545,
"grad_norm": 1.1019212007522583,
"learning_rate": 3.0037735849056606e-06,
"loss": 0.1604,
"step": 677
},
{
"epoch": 1.3815588385124808,
"grad_norm": 1.262918472290039,
"learning_rate": 3e-06,
"loss": 0.1713,
"step": 678
},
{
"epoch": 1.3835965359144167,
"grad_norm": 1.1134512424468994,
"learning_rate": 2.99622641509434e-06,
"loss": 0.1738,
"step": 679
},
{
"epoch": 1.3856342333163525,
"grad_norm": 1.1910215616226196,
"learning_rate": 2.9924528301886795e-06,
"loss": 0.1854,
"step": 680
},
{
"epoch": 1.3876719307182883,
"grad_norm": 1.0705041885375977,
"learning_rate": 2.9886792452830194e-06,
"loss": 0.1679,
"step": 681
},
{
"epoch": 1.3897096281202241,
"grad_norm": 1.0849546194076538,
"learning_rate": 2.9849056603773585e-06,
"loss": 0.1732,
"step": 682
},
{
"epoch": 1.39174732552216,
"grad_norm": 1.1088389158248901,
"learning_rate": 2.9811320754716984e-06,
"loss": 0.1832,
"step": 683
},
{
"epoch": 1.3937850229240958,
"grad_norm": 1.1701173782348633,
"learning_rate": 2.977358490566038e-06,
"loss": 0.1832,
"step": 684
},
{
"epoch": 1.3958227203260316,
"grad_norm": 1.1918519735336304,
"learning_rate": 2.9735849056603778e-06,
"loss": 0.1863,
"step": 685
},
{
"epoch": 1.3978604177279674,
"grad_norm": 1.207116723060608,
"learning_rate": 2.9698113207547173e-06,
"loss": 0.1806,
"step": 686
},
{
"epoch": 1.3998981151299033,
"grad_norm": 1.2102634906768799,
"learning_rate": 2.9660377358490568e-06,
"loss": 0.1759,
"step": 687
},
{
"epoch": 1.401935812531839,
"grad_norm": 1.1316732168197632,
"learning_rate": 2.9622641509433963e-06,
"loss": 0.17,
"step": 688
},
{
"epoch": 1.403973509933775,
"grad_norm": 1.204567790031433,
"learning_rate": 2.958490566037736e-06,
"loss": 0.1908,
"step": 689
},
{
"epoch": 1.4060112073357107,
"grad_norm": 1.0931925773620605,
"learning_rate": 2.9547169811320757e-06,
"loss": 0.1784,
"step": 690
},
{
"epoch": 1.4080489047376465,
"grad_norm": 1.2366472482681274,
"learning_rate": 2.9509433962264156e-06,
"loss": 0.2053,
"step": 691
},
{
"epoch": 1.4100866021395824,
"grad_norm": 1.169756531715393,
"learning_rate": 2.9471698113207546e-06,
"loss": 0.1803,
"step": 692
},
{
"epoch": 1.4121242995415182,
"grad_norm": 1.271429419517517,
"learning_rate": 2.9433962264150946e-06,
"loss": 0.1867,
"step": 693
},
{
"epoch": 1.414161996943454,
"grad_norm": 1.2226650714874268,
"learning_rate": 2.939622641509434e-06,
"loss": 0.1643,
"step": 694
},
{
"epoch": 1.4161996943453898,
"grad_norm": 1.2417409420013428,
"learning_rate": 2.935849056603774e-06,
"loss": 0.1897,
"step": 695
},
{
"epoch": 1.4182373917473257,
"grad_norm": 1.24673593044281,
"learning_rate": 2.932075471698114e-06,
"loss": 0.1648,
"step": 696
},
{
"epoch": 1.4202750891492613,
"grad_norm": 1.336515188217163,
"learning_rate": 2.928301886792453e-06,
"loss": 0.1913,
"step": 697
},
{
"epoch": 1.422312786551197,
"grad_norm": 1.1495544910430908,
"learning_rate": 2.9245283018867924e-06,
"loss": 0.1825,
"step": 698
},
{
"epoch": 1.424350483953133,
"grad_norm": 1.181207537651062,
"learning_rate": 2.9207547169811324e-06,
"loss": 0.1814,
"step": 699
},
{
"epoch": 1.4263881813550687,
"grad_norm": 1.2883107662200928,
"learning_rate": 2.9169811320754723e-06,
"loss": 0.1859,
"step": 700
},
{
"epoch": 1.4284258787570046,
"grad_norm": 1.14235520362854,
"learning_rate": 2.9132075471698118e-06,
"loss": 0.1846,
"step": 701
},
{
"epoch": 1.4304635761589404,
"grad_norm": 1.0994147062301636,
"learning_rate": 2.909433962264151e-06,
"loss": 0.18,
"step": 702
},
{
"epoch": 1.4325012735608762,
"grad_norm": 1.2511541843414307,
"learning_rate": 2.9056603773584907e-06,
"loss": 0.1759,
"step": 703
},
{
"epoch": 1.434538970962812,
"grad_norm": 1.0954980850219727,
"learning_rate": 2.9018867924528307e-06,
"loss": 0.1724,
"step": 704
},
{
"epoch": 1.4365766683647478,
"grad_norm": 1.3084522485733032,
"learning_rate": 2.89811320754717e-06,
"loss": 0.179,
"step": 705
},
{
"epoch": 1.4386143657666837,
"grad_norm": 1.1592984199523926,
"learning_rate": 2.89433962264151e-06,
"loss": 0.1798,
"step": 706
},
{
"epoch": 1.4406520631686195,
"grad_norm": 1.1409646272659302,
"learning_rate": 2.890566037735849e-06,
"loss": 0.175,
"step": 707
},
{
"epoch": 1.4426897605705553,
"grad_norm": 1.3026984930038452,
"learning_rate": 2.886792452830189e-06,
"loss": 0.1801,
"step": 708
},
{
"epoch": 1.4447274579724911,
"grad_norm": 1.115729570388794,
"learning_rate": 2.8830188679245285e-06,
"loss": 0.188,
"step": 709
},
{
"epoch": 1.446765155374427,
"grad_norm": 1.3142112493515015,
"learning_rate": 2.8792452830188684e-06,
"loss": 0.187,
"step": 710
},
{
"epoch": 1.4488028527763628,
"grad_norm": 1.2339842319488525,
"learning_rate": 2.875471698113208e-06,
"loss": 0.1614,
"step": 711
},
{
"epoch": 1.4508405501782986,
"grad_norm": 1.2981687784194946,
"learning_rate": 2.871698113207547e-06,
"loss": 0.1688,
"step": 712
},
{
"epoch": 1.4528782475802342,
"grad_norm": 1.1264586448669434,
"learning_rate": 2.867924528301887e-06,
"loss": 0.1754,
"step": 713
},
{
"epoch": 1.45491594498217,
"grad_norm": 1.1794300079345703,
"learning_rate": 2.864150943396227e-06,
"loss": 0.1876,
"step": 714
},
{
"epoch": 1.4569536423841059,
"grad_norm": 1.0934234857559204,
"learning_rate": 2.8603773584905663e-06,
"loss": 0.18,
"step": 715
},
{
"epoch": 1.4589913397860417,
"grad_norm": 1.1383419036865234,
"learning_rate": 2.8566037735849062e-06,
"loss": 0.1812,
"step": 716
},
{
"epoch": 1.4610290371879775,
"grad_norm": 1.1334176063537598,
"learning_rate": 2.8528301886792453e-06,
"loss": 0.1699,
"step": 717
},
{
"epoch": 1.4630667345899133,
"grad_norm": 1.2105752229690552,
"learning_rate": 2.8490566037735852e-06,
"loss": 0.1886,
"step": 718
},
{
"epoch": 1.4651044319918491,
"grad_norm": 1.1222751140594482,
"learning_rate": 2.8452830188679247e-06,
"loss": 0.1677,
"step": 719
},
{
"epoch": 1.467142129393785,
"grad_norm": 1.0429809093475342,
"learning_rate": 2.8415094339622646e-06,
"loss": 0.1801,
"step": 720
},
{
"epoch": 1.4691798267957208,
"grad_norm": 1.1673039197921753,
"learning_rate": 2.837735849056604e-06,
"loss": 0.1824,
"step": 721
},
{
"epoch": 1.4712175241976566,
"grad_norm": 1.2965126037597656,
"learning_rate": 2.8339622641509436e-06,
"loss": 0.1789,
"step": 722
},
{
"epoch": 1.4732552215995924,
"grad_norm": 1.1965491771697998,
"learning_rate": 2.830188679245283e-06,
"loss": 0.1875,
"step": 723
},
{
"epoch": 1.4752929190015283,
"grad_norm": 1.1529309749603271,
"learning_rate": 2.826415094339623e-06,
"loss": 0.1846,
"step": 724
},
{
"epoch": 1.477330616403464,
"grad_norm": 1.1195148229599,
"learning_rate": 2.8226415094339625e-06,
"loss": 0.1655,
"step": 725
},
{
"epoch": 1.4793683138054,
"grad_norm": 1.2534137964248657,
"learning_rate": 2.8188679245283024e-06,
"loss": 0.1779,
"step": 726
},
{
"epoch": 1.4814060112073357,
"grad_norm": 1.1430234909057617,
"learning_rate": 2.8150943396226415e-06,
"loss": 0.1855,
"step": 727
},
{
"epoch": 1.4834437086092715,
"grad_norm": 1.1733477115631104,
"learning_rate": 2.8113207547169814e-06,
"loss": 0.1818,
"step": 728
},
{
"epoch": 1.4854814060112074,
"grad_norm": 1.2729791402816772,
"learning_rate": 2.807547169811321e-06,
"loss": 0.1845,
"step": 729
},
{
"epoch": 1.4875191034131432,
"grad_norm": 1.2047133445739746,
"learning_rate": 2.803773584905661e-06,
"loss": 0.1853,
"step": 730
},
{
"epoch": 1.489556800815079,
"grad_norm": 1.0154218673706055,
"learning_rate": 2.8000000000000003e-06,
"loss": 0.1845,
"step": 731
},
{
"epoch": 1.4915944982170148,
"grad_norm": 1.0939674377441406,
"learning_rate": 2.7962264150943398e-06,
"loss": 0.1856,
"step": 732
},
{
"epoch": 1.4936321956189507,
"grad_norm": 1.1324870586395264,
"learning_rate": 2.7924528301886793e-06,
"loss": 0.1755,
"step": 733
},
{
"epoch": 1.4956698930208865,
"grad_norm": 1.4036580324172974,
"learning_rate": 2.788679245283019e-06,
"loss": 0.2023,
"step": 734
},
{
"epoch": 1.4977075904228223,
"grad_norm": 1.1628963947296143,
"learning_rate": 2.7849056603773587e-06,
"loss": 0.1787,
"step": 735
},
{
"epoch": 1.4997452878247581,
"grad_norm": 1.0612685680389404,
"learning_rate": 2.7811320754716986e-06,
"loss": 0.1709,
"step": 736
},
{
"epoch": 1.501782985226694,
"grad_norm": 1.1758002042770386,
"learning_rate": 2.7773584905660377e-06,
"loss": 0.1844,
"step": 737
},
{
"epoch": 1.5038206826286298,
"grad_norm": 1.1747825145721436,
"learning_rate": 2.7735849056603776e-06,
"loss": 0.1759,
"step": 738
},
{
"epoch": 1.5058583800305656,
"grad_norm": 1.1918827295303345,
"learning_rate": 2.769811320754717e-06,
"loss": 0.1685,
"step": 739
},
{
"epoch": 1.5078960774325014,
"grad_norm": 1.1047258377075195,
"learning_rate": 2.766037735849057e-06,
"loss": 0.1826,
"step": 740
},
{
"epoch": 1.5099337748344372,
"grad_norm": 1.209409236907959,
"learning_rate": 2.762264150943397e-06,
"loss": 0.1703,
"step": 741
},
{
"epoch": 1.5119714722363728,
"grad_norm": 1.1031354665756226,
"learning_rate": 2.758490566037736e-06,
"loss": 0.1766,
"step": 742
},
{
"epoch": 1.5140091696383087,
"grad_norm": 1.2434014081954956,
"learning_rate": 2.7547169811320755e-06,
"loss": 0.1844,
"step": 743
},
{
"epoch": 1.5160468670402445,
"grad_norm": 1.177281379699707,
"learning_rate": 2.7509433962264154e-06,
"loss": 0.186,
"step": 744
},
{
"epoch": 1.5180845644421803,
"grad_norm": 1.0548818111419678,
"learning_rate": 2.7471698113207553e-06,
"loss": 0.1675,
"step": 745
},
{
"epoch": 1.5201222618441161,
"grad_norm": 1.1306318044662476,
"learning_rate": 2.7433962264150944e-06,
"loss": 0.1713,
"step": 746
},
{
"epoch": 1.522159959246052,
"grad_norm": 1.205263376235962,
"learning_rate": 2.739622641509434e-06,
"loss": 0.1906,
"step": 747
},
{
"epoch": 1.5241976566479878,
"grad_norm": 2.5892493724823,
"learning_rate": 2.7358490566037738e-06,
"loss": 0.1757,
"step": 748
},
{
"epoch": 1.5262353540499236,
"grad_norm": 1.0715084075927734,
"learning_rate": 2.7320754716981137e-06,
"loss": 0.173,
"step": 749
},
{
"epoch": 1.5282730514518594,
"grad_norm": 1.231529712677002,
"learning_rate": 2.728301886792453e-06,
"loss": 0.1701,
"step": 750
},
{
"epoch": 1.5303107488537953,
"grad_norm": 1.2592768669128418,
"learning_rate": 2.7245283018867922e-06,
"loss": 0.1774,
"step": 751
},
{
"epoch": 1.532348446255731,
"grad_norm": 1.2342033386230469,
"learning_rate": 2.720754716981132e-06,
"loss": 0.17,
"step": 752
},
{
"epoch": 1.5343861436576667,
"grad_norm": 1.1225703954696655,
"learning_rate": 2.716981132075472e-06,
"loss": 0.1786,
"step": 753
},
{
"epoch": 1.5364238410596025,
"grad_norm": 1.204437494277954,
"learning_rate": 2.7132075471698116e-06,
"loss": 0.182,
"step": 754
},
{
"epoch": 1.5384615384615383,
"grad_norm": 1.152274489402771,
"learning_rate": 2.7094339622641515e-06,
"loss": 0.1845,
"step": 755
},
{
"epoch": 1.5404992358634741,
"grad_norm": 1.268399715423584,
"learning_rate": 2.7056603773584905e-06,
"loss": 0.1866,
"step": 756
},
{
"epoch": 1.54253693326541,
"grad_norm": 1.3325903415679932,
"learning_rate": 2.7018867924528304e-06,
"loss": 0.1788,
"step": 757
},
{
"epoch": 1.5445746306673458,
"grad_norm": 1.164884090423584,
"learning_rate": 2.69811320754717e-06,
"loss": 0.1863,
"step": 758
},
{
"epoch": 1.5466123280692816,
"grad_norm": 1.1347957849502563,
"learning_rate": 2.69433962264151e-06,
"loss": 0.1902,
"step": 759
},
{
"epoch": 1.5486500254712174,
"grad_norm": 1.1705092191696167,
"learning_rate": 2.6905660377358493e-06,
"loss": 0.1757,
"step": 760
},
{
"epoch": 1.5506877228731533,
"grad_norm": 1.1735482215881348,
"learning_rate": 2.686792452830189e-06,
"loss": 0.1743,
"step": 761
},
{
"epoch": 1.552725420275089,
"grad_norm": 1.1496127843856812,
"learning_rate": 2.6830188679245283e-06,
"loss": 0.1704,
"step": 762
},
{
"epoch": 1.554763117677025,
"grad_norm": 1.1327245235443115,
"learning_rate": 2.6792452830188682e-06,
"loss": 0.1687,
"step": 763
},
{
"epoch": 1.5568008150789607,
"grad_norm": 1.235737919807434,
"learning_rate": 2.6754716981132077e-06,
"loss": 0.1699,
"step": 764
},
{
"epoch": 1.5588385124808966,
"grad_norm": 1.0961453914642334,
"learning_rate": 2.6716981132075476e-06,
"loss": 0.174,
"step": 765
},
{
"epoch": 1.5608762098828324,
"grad_norm": 1.1706377267837524,
"learning_rate": 2.6679245283018867e-06,
"loss": 0.1679,
"step": 766
},
{
"epoch": 1.5629139072847682,
"grad_norm": 1.314253330230713,
"learning_rate": 2.6641509433962266e-06,
"loss": 0.1859,
"step": 767
},
{
"epoch": 1.564951604686704,
"grad_norm": 1.0271321535110474,
"learning_rate": 2.660377358490566e-06,
"loss": 0.1717,
"step": 768
},
{
"epoch": 1.5669893020886398,
"grad_norm": 1.11105215549469,
"learning_rate": 2.656603773584906e-06,
"loss": 0.1699,
"step": 769
},
{
"epoch": 1.5690269994905757,
"grad_norm": 1.2342256307601929,
"learning_rate": 2.6528301886792455e-06,
"loss": 0.1836,
"step": 770
},
{
"epoch": 1.5710646968925115,
"grad_norm": 1.208130121231079,
"learning_rate": 2.649056603773585e-06,
"loss": 0.1708,
"step": 771
},
{
"epoch": 1.5731023942944473,
"grad_norm": 1.235351324081421,
"learning_rate": 2.6452830188679245e-06,
"loss": 0.1976,
"step": 772
},
{
"epoch": 1.5751400916963831,
"grad_norm": 1.0710421800613403,
"learning_rate": 2.6415094339622644e-06,
"loss": 0.1734,
"step": 773
},
{
"epoch": 1.577177789098319,
"grad_norm": 0.9788026213645935,
"learning_rate": 2.637735849056604e-06,
"loss": 0.1701,
"step": 774
},
{
"epoch": 1.5792154865002548,
"grad_norm": 1.1931087970733643,
"learning_rate": 2.633962264150944e-06,
"loss": 0.1667,
"step": 775
},
{
"epoch": 1.5812531839021906,
"grad_norm": 1.242144227027893,
"learning_rate": 2.630188679245283e-06,
"loss": 0.1923,
"step": 776
},
{
"epoch": 1.5832908813041264,
"grad_norm": 1.2944048643112183,
"learning_rate": 2.626415094339623e-06,
"loss": 0.1768,
"step": 777
},
{
"epoch": 1.5853285787060623,
"grad_norm": 1.0808852910995483,
"learning_rate": 2.6226415094339623e-06,
"loss": 0.1721,
"step": 778
},
{
"epoch": 1.587366276107998,
"grad_norm": 1.147532343864441,
"learning_rate": 2.6188679245283022e-06,
"loss": 0.1618,
"step": 779
},
{
"epoch": 1.589403973509934,
"grad_norm": 1.2777063846588135,
"learning_rate": 2.615094339622642e-06,
"loss": 0.1831,
"step": 780
},
{
"epoch": 1.5914416709118697,
"grad_norm": 1.1522384881973267,
"learning_rate": 2.611320754716981e-06,
"loss": 0.1725,
"step": 781
},
{
"epoch": 1.5934793683138055,
"grad_norm": 1.1645333766937256,
"learning_rate": 2.6075471698113207e-06,
"loss": 0.1724,
"step": 782
},
{
"epoch": 1.5955170657157414,
"grad_norm": 1.1945953369140625,
"learning_rate": 2.6037735849056606e-06,
"loss": 0.182,
"step": 783
},
{
"epoch": 1.5975547631176772,
"grad_norm": 1.2776046991348267,
"learning_rate": 2.6e-06,
"loss": 0.1783,
"step": 784
},
{
"epoch": 1.5995924605196128,
"grad_norm": 1.0407108068466187,
"learning_rate": 2.59622641509434e-06,
"loss": 0.1651,
"step": 785
},
{
"epoch": 1.6016301579215486,
"grad_norm": 1.1741459369659424,
"learning_rate": 2.592452830188679e-06,
"loss": 0.1759,
"step": 786
},
{
"epoch": 1.6036678553234844,
"grad_norm": 1.1339528560638428,
"learning_rate": 2.588679245283019e-06,
"loss": 0.1753,
"step": 787
},
{
"epoch": 1.6057055527254203,
"grad_norm": 1.5073323249816895,
"learning_rate": 2.5849056603773585e-06,
"loss": 0.1828,
"step": 788
},
{
"epoch": 1.607743250127356,
"grad_norm": 1.097970962524414,
"learning_rate": 2.5811320754716984e-06,
"loss": 0.1709,
"step": 789
},
{
"epoch": 1.609780947529292,
"grad_norm": 1.0759773254394531,
"learning_rate": 2.5773584905660383e-06,
"loss": 0.1562,
"step": 790
},
{
"epoch": 1.6118186449312277,
"grad_norm": 1.1199358701705933,
"learning_rate": 2.5735849056603774e-06,
"loss": 0.1751,
"step": 791
},
{
"epoch": 1.6138563423331636,
"grad_norm": 1.162474513053894,
"learning_rate": 2.569811320754717e-06,
"loss": 0.1691,
"step": 792
},
{
"epoch": 1.6158940397350994,
"grad_norm": 1.170835256576538,
"learning_rate": 2.5660377358490568e-06,
"loss": 0.179,
"step": 793
},
{
"epoch": 1.6179317371370352,
"grad_norm": 1.087983250617981,
"learning_rate": 2.5622641509433967e-06,
"loss": 0.1736,
"step": 794
},
{
"epoch": 1.6199694345389708,
"grad_norm": 1.1620844602584839,
"learning_rate": 2.558490566037736e-06,
"loss": 0.1815,
"step": 795
},
{
"epoch": 1.6220071319409066,
"grad_norm": 1.1823047399520874,
"learning_rate": 2.5547169811320753e-06,
"loss": 0.1797,
"step": 796
},
{
"epoch": 1.6240448293428424,
"grad_norm": 1.1422289609909058,
"learning_rate": 2.550943396226415e-06,
"loss": 0.1812,
"step": 797
},
{
"epoch": 1.6260825267447783,
"grad_norm": 1.2025611400604248,
"learning_rate": 2.547169811320755e-06,
"loss": 0.1807,
"step": 798
},
{
"epoch": 1.628120224146714,
"grad_norm": 1.140370488166809,
"learning_rate": 2.5433962264150946e-06,
"loss": 0.1782,
"step": 799
},
{
"epoch": 1.63015792154865,
"grad_norm": 1.1452966928482056,
"learning_rate": 2.5396226415094345e-06,
"loss": 0.1724,
"step": 800
},
{
"epoch": 1.6321956189505857,
"grad_norm": 1.217185616493225,
"learning_rate": 2.5358490566037736e-06,
"loss": 0.1807,
"step": 801
},
{
"epoch": 1.6342333163525216,
"grad_norm": 1.0574156045913696,
"learning_rate": 2.5320754716981135e-06,
"loss": 0.1694,
"step": 802
},
{
"epoch": 1.6362710137544574,
"grad_norm": 1.015283226966858,
"learning_rate": 2.528301886792453e-06,
"loss": 0.1713,
"step": 803
},
{
"epoch": 1.6383087111563932,
"grad_norm": 1.1992040872573853,
"learning_rate": 2.524528301886793e-06,
"loss": 0.1844,
"step": 804
},
{
"epoch": 1.640346408558329,
"grad_norm": 1.2918540239334106,
"learning_rate": 2.5207547169811324e-06,
"loss": 0.1813,
"step": 805
},
{
"epoch": 1.6423841059602649,
"grad_norm": 1.1141362190246582,
"learning_rate": 2.516981132075472e-06,
"loss": 0.1763,
"step": 806
},
{
"epoch": 1.6444218033622007,
"grad_norm": 1.0930787324905396,
"learning_rate": 2.5132075471698114e-06,
"loss": 0.184,
"step": 807
},
{
"epoch": 1.6464595007641365,
"grad_norm": 1.1243940591812134,
"learning_rate": 2.5094339622641513e-06,
"loss": 0.1821,
"step": 808
},
{
"epoch": 1.6484971981660723,
"grad_norm": 1.1842948198318481,
"learning_rate": 2.5056603773584908e-06,
"loss": 0.1684,
"step": 809
},
{
"epoch": 1.6505348955680081,
"grad_norm": 1.2824788093566895,
"learning_rate": 2.5018867924528307e-06,
"loss": 0.182,
"step": 810
},
{
"epoch": 1.652572592969944,
"grad_norm": 1.1476082801818848,
"learning_rate": 2.49811320754717e-06,
"loss": 0.1847,
"step": 811
},
{
"epoch": 1.6546102903718798,
"grad_norm": 1.1569533348083496,
"learning_rate": 2.4943396226415097e-06,
"loss": 0.1815,
"step": 812
},
{
"epoch": 1.6566479877738156,
"grad_norm": 1.1782304048538208,
"learning_rate": 2.490566037735849e-06,
"loss": 0.1754,
"step": 813
},
{
"epoch": 1.6586856851757514,
"grad_norm": 1.1351999044418335,
"learning_rate": 2.486792452830189e-06,
"loss": 0.189,
"step": 814
},
{
"epoch": 1.6607233825776873,
"grad_norm": 1.1230946779251099,
"learning_rate": 2.4830188679245285e-06,
"loss": 0.1781,
"step": 815
},
{
"epoch": 1.662761079979623,
"grad_norm": 1.062568187713623,
"learning_rate": 2.479245283018868e-06,
"loss": 0.1665,
"step": 816
},
{
"epoch": 1.664798777381559,
"grad_norm": 1.1602753400802612,
"learning_rate": 2.4754716981132075e-06,
"loss": 0.1738,
"step": 817
},
{
"epoch": 1.6668364747834947,
"grad_norm": 1.1816747188568115,
"learning_rate": 2.4716981132075474e-06,
"loss": 0.1675,
"step": 818
},
{
"epoch": 1.6688741721854305,
"grad_norm": 1.1682571172714233,
"learning_rate": 2.467924528301887e-06,
"loss": 0.1657,
"step": 819
},
{
"epoch": 1.6709118695873664,
"grad_norm": 1.0787543058395386,
"learning_rate": 2.4641509433962264e-06,
"loss": 0.1717,
"step": 820
},
{
"epoch": 1.6729495669893022,
"grad_norm": 1.1307450532913208,
"learning_rate": 2.4603773584905663e-06,
"loss": 0.1708,
"step": 821
},
{
"epoch": 1.674987264391238,
"grad_norm": 1.1192117929458618,
"learning_rate": 2.456603773584906e-06,
"loss": 0.1799,
"step": 822
},
{
"epoch": 1.6770249617931738,
"grad_norm": 1.1960910558700562,
"learning_rate": 2.4528301886792453e-06,
"loss": 0.1703,
"step": 823
},
{
"epoch": 1.6790626591951097,
"grad_norm": 1.1331156492233276,
"learning_rate": 2.4490566037735852e-06,
"loss": 0.1593,
"step": 824
},
{
"epoch": 1.6811003565970455,
"grad_norm": 1.2134394645690918,
"learning_rate": 2.4452830188679247e-06,
"loss": 0.1964,
"step": 825
},
{
"epoch": 1.6831380539989813,
"grad_norm": 1.178653597831726,
"learning_rate": 2.4415094339622642e-06,
"loss": 0.1864,
"step": 826
},
{
"epoch": 1.685175751400917,
"grad_norm": 1.0972850322723389,
"learning_rate": 2.4377358490566037e-06,
"loss": 0.1637,
"step": 827
},
{
"epoch": 1.6872134488028527,
"grad_norm": 1.0110701322555542,
"learning_rate": 2.4339622641509436e-06,
"loss": 0.1758,
"step": 828
},
{
"epoch": 1.6892511462047886,
"grad_norm": 1.0509426593780518,
"learning_rate": 2.4301886792452835e-06,
"loss": 0.1708,
"step": 829
},
{
"epoch": 1.6912888436067244,
"grad_norm": 1.247532844543457,
"learning_rate": 2.4264150943396226e-06,
"loss": 0.1754,
"step": 830
},
{
"epoch": 1.6933265410086602,
"grad_norm": 1.1235079765319824,
"learning_rate": 2.4226415094339625e-06,
"loss": 0.1699,
"step": 831
},
{
"epoch": 1.695364238410596,
"grad_norm": 1.2192139625549316,
"learning_rate": 2.418867924528302e-06,
"loss": 0.1851,
"step": 832
},
{
"epoch": 1.6974019358125318,
"grad_norm": 1.2487667798995972,
"learning_rate": 2.415094339622642e-06,
"loss": 0.1926,
"step": 833
},
{
"epoch": 1.6994396332144677,
"grad_norm": 1.1878374814987183,
"learning_rate": 2.4113207547169814e-06,
"loss": 0.2017,
"step": 834
},
{
"epoch": 1.7014773306164035,
"grad_norm": 1.027300238609314,
"learning_rate": 2.407547169811321e-06,
"loss": 0.1725,
"step": 835
},
{
"epoch": 1.7035150280183393,
"grad_norm": 1.0987987518310547,
"learning_rate": 2.403773584905661e-06,
"loss": 0.1793,
"step": 836
},
{
"epoch": 1.7055527254202751,
"grad_norm": 1.08310067653656,
"learning_rate": 2.4000000000000003e-06,
"loss": 0.1715,
"step": 837
},
{
"epoch": 1.7075904228222107,
"grad_norm": 1.255993366241455,
"learning_rate": 2.39622641509434e-06,
"loss": 0.1769,
"step": 838
},
{
"epoch": 1.7096281202241466,
"grad_norm": 1.1966819763183594,
"learning_rate": 2.3924528301886797e-06,
"loss": 0.1661,
"step": 839
},
{
"epoch": 1.7116658176260824,
"grad_norm": 1.22041916847229,
"learning_rate": 2.388679245283019e-06,
"loss": 0.172,
"step": 840
},
{
"epoch": 1.7137035150280182,
"grad_norm": 1.0473703145980835,
"learning_rate": 2.3849056603773587e-06,
"loss": 0.1555,
"step": 841
},
{
"epoch": 1.715741212429954,
"grad_norm": 1.0921486616134644,
"learning_rate": 2.381132075471698e-06,
"loss": 0.1739,
"step": 842
},
{
"epoch": 1.7177789098318899,
"grad_norm": 1.1403447389602661,
"learning_rate": 2.377358490566038e-06,
"loss": 0.1807,
"step": 843
},
{
"epoch": 1.7198166072338257,
"grad_norm": 1.1131690740585327,
"learning_rate": 2.3735849056603776e-06,
"loss": 0.1874,
"step": 844
},
{
"epoch": 1.7218543046357615,
"grad_norm": 1.1460295915603638,
"learning_rate": 2.369811320754717e-06,
"loss": 0.1709,
"step": 845
},
{
"epoch": 1.7238920020376973,
"grad_norm": 1.1869096755981445,
"learning_rate": 2.366037735849057e-06,
"loss": 0.1883,
"step": 846
},
{
"epoch": 1.7259296994396331,
"grad_norm": 1.1736819744110107,
"learning_rate": 2.3622641509433965e-06,
"loss": 0.1754,
"step": 847
},
{
"epoch": 1.727967396841569,
"grad_norm": 1.213629126548767,
"learning_rate": 2.358490566037736e-06,
"loss": 0.1677,
"step": 848
},
{
"epoch": 1.7300050942435048,
"grad_norm": 1.0772464275360107,
"learning_rate": 2.3547169811320755e-06,
"loss": 0.1769,
"step": 849
},
{
"epoch": 1.7320427916454406,
"grad_norm": 1.1553244590759277,
"learning_rate": 2.3509433962264154e-06,
"loss": 0.1817,
"step": 850
},
{
"epoch": 1.7340804890473764,
"grad_norm": 1.0742902755737305,
"learning_rate": 2.347169811320755e-06,
"loss": 0.1635,
"step": 851
},
{
"epoch": 1.7361181864493123,
"grad_norm": 1.1944258213043213,
"learning_rate": 2.3433962264150944e-06,
"loss": 0.1757,
"step": 852
},
{
"epoch": 1.738155883851248,
"grad_norm": 1.1923333406448364,
"learning_rate": 2.3396226415094343e-06,
"loss": 0.1665,
"step": 853
},
{
"epoch": 1.740193581253184,
"grad_norm": 1.086665153503418,
"learning_rate": 2.3358490566037738e-06,
"loss": 0.1803,
"step": 854
},
{
"epoch": 1.7422312786551197,
"grad_norm": 1.0686219930648804,
"learning_rate": 2.3320754716981133e-06,
"loss": 0.1757,
"step": 855
},
{
"epoch": 1.7442689760570556,
"grad_norm": 1.6613824367523193,
"learning_rate": 2.328301886792453e-06,
"loss": 0.1927,
"step": 856
},
{
"epoch": 1.7463066734589914,
"grad_norm": 1.305106282234192,
"learning_rate": 2.3245283018867927e-06,
"loss": 0.1882,
"step": 857
},
{
"epoch": 1.7483443708609272,
"grad_norm": 1.091124176979065,
"learning_rate": 2.320754716981132e-06,
"loss": 0.1707,
"step": 858
},
{
"epoch": 1.750382068262863,
"grad_norm": 1.073729157447815,
"learning_rate": 2.3169811320754717e-06,
"loss": 0.1891,
"step": 859
},
{
"epoch": 1.7524197656647988,
"grad_norm": 1.2147339582443237,
"learning_rate": 2.3132075471698116e-06,
"loss": 0.1734,
"step": 860
},
{
"epoch": 1.7544574630667347,
"grad_norm": 1.085634708404541,
"learning_rate": 2.309433962264151e-06,
"loss": 0.1577,
"step": 861
},
{
"epoch": 1.7564951604686705,
"grad_norm": 1.220919132232666,
"learning_rate": 2.3056603773584906e-06,
"loss": 0.1763,
"step": 862
},
{
"epoch": 1.7585328578706063,
"grad_norm": 1.3067682981491089,
"learning_rate": 2.3018867924528305e-06,
"loss": 0.1805,
"step": 863
},
{
"epoch": 1.7605705552725421,
"grad_norm": 1.3163460493087769,
"learning_rate": 2.29811320754717e-06,
"loss": 0.1737,
"step": 864
},
{
"epoch": 1.762608252674478,
"grad_norm": 1.1450026035308838,
"learning_rate": 2.2943396226415095e-06,
"loss": 0.1727,
"step": 865
},
{
"epoch": 1.7646459500764138,
"grad_norm": 1.0936638116836548,
"learning_rate": 2.2905660377358494e-06,
"loss": 0.1772,
"step": 866
},
{
"epoch": 1.7666836474783496,
"grad_norm": 1.2066489458084106,
"learning_rate": 2.286792452830189e-06,
"loss": 0.1777,
"step": 867
},
{
"epoch": 1.7687213448802854,
"grad_norm": 1.2631739377975464,
"learning_rate": 2.2830188679245283e-06,
"loss": 0.1765,
"step": 868
},
{
"epoch": 1.7707590422822213,
"grad_norm": 1.1708970069885254,
"learning_rate": 2.279245283018868e-06,
"loss": 0.1834,
"step": 869
},
{
"epoch": 1.7727967396841569,
"grad_norm": 1.0745712518692017,
"learning_rate": 2.2754716981132078e-06,
"loss": 0.1659,
"step": 870
},
{
"epoch": 1.7748344370860927,
"grad_norm": 1.243639588356018,
"learning_rate": 2.2716981132075477e-06,
"loss": 0.1683,
"step": 871
},
{
"epoch": 1.7768721344880285,
"grad_norm": 1.2835688591003418,
"learning_rate": 2.2679245283018867e-06,
"loss": 0.1749,
"step": 872
},
{
"epoch": 1.7789098318899643,
"grad_norm": 1.3315813541412354,
"learning_rate": 2.2641509433962266e-06,
"loss": 0.187,
"step": 873
},
{
"epoch": 1.7809475292919001,
"grad_norm": 1.525321125984192,
"learning_rate": 2.260377358490566e-06,
"loss": 0.1838,
"step": 874
},
{
"epoch": 1.782985226693836,
"grad_norm": 1.1951662302017212,
"learning_rate": 2.256603773584906e-06,
"loss": 0.1902,
"step": 875
},
{
"epoch": 1.7850229240957718,
"grad_norm": 1.2421764135360718,
"learning_rate": 2.2528301886792455e-06,
"loss": 0.1825,
"step": 876
},
{
"epoch": 1.7870606214977076,
"grad_norm": 1.1425124406814575,
"learning_rate": 2.249056603773585e-06,
"loss": 0.1647,
"step": 877
},
{
"epoch": 1.7890983188996434,
"grad_norm": 1.6294941902160645,
"learning_rate": 2.245283018867925e-06,
"loss": 0.1848,
"step": 878
},
{
"epoch": 1.7911360163015793,
"grad_norm": 1.0840221643447876,
"learning_rate": 2.241509433962264e-06,
"loss": 0.178,
"step": 879
},
{
"epoch": 1.793173713703515,
"grad_norm": 1.1134402751922607,
"learning_rate": 2.237735849056604e-06,
"loss": 0.1563,
"step": 880
},
{
"epoch": 1.7952114111054507,
"grad_norm": 1.2107329368591309,
"learning_rate": 2.233962264150944e-06,
"loss": 0.1756,
"step": 881
},
{
"epoch": 1.7972491085073865,
"grad_norm": 1.2982094287872314,
"learning_rate": 2.2301886792452833e-06,
"loss": 0.1793,
"step": 882
},
{
"epoch": 1.7992868059093223,
"grad_norm": 1.2917886972427368,
"learning_rate": 2.226415094339623e-06,
"loss": 0.1666,
"step": 883
},
{
"epoch": 1.8013245033112582,
"grad_norm": 1.23494553565979,
"learning_rate": 2.2226415094339623e-06,
"loss": 0.1707,
"step": 884
},
{
"epoch": 1.803362200713194,
"grad_norm": 1.1923739910125732,
"learning_rate": 2.2188679245283022e-06,
"loss": 0.1767,
"step": 885
},
{
"epoch": 1.8053998981151298,
"grad_norm": 1.1137254238128662,
"learning_rate": 2.2150943396226417e-06,
"loss": 0.1795,
"step": 886
},
{
"epoch": 1.8074375955170656,
"grad_norm": 1.1190637350082397,
"learning_rate": 2.2113207547169812e-06,
"loss": 0.1766,
"step": 887
},
{
"epoch": 1.8094752929190014,
"grad_norm": 1.1797064542770386,
"learning_rate": 2.207547169811321e-06,
"loss": 0.1713,
"step": 888
},
{
"epoch": 1.8115129903209373,
"grad_norm": 1.1107820272445679,
"learning_rate": 2.2037735849056606e-06,
"loss": 0.1826,
"step": 889
},
{
"epoch": 1.813550687722873,
"grad_norm": 1.1796709299087524,
"learning_rate": 2.2e-06,
"loss": 0.1771,
"step": 890
},
{
"epoch": 1.815588385124809,
"grad_norm": 1.0448757410049438,
"learning_rate": 2.19622641509434e-06,
"loss": 0.1673,
"step": 891
},
{
"epoch": 1.8176260825267447,
"grad_norm": 1.1002962589263916,
"learning_rate": 2.1924528301886795e-06,
"loss": 0.1612,
"step": 892
},
{
"epoch": 1.8196637799286806,
"grad_norm": 1.2181810140609741,
"learning_rate": 2.188679245283019e-06,
"loss": 0.1752,
"step": 893
},
{
"epoch": 1.8217014773306164,
"grad_norm": 1.2177342176437378,
"learning_rate": 2.1849056603773585e-06,
"loss": 0.1713,
"step": 894
},
{
"epoch": 1.8237391747325522,
"grad_norm": 1.070660948753357,
"learning_rate": 2.1811320754716984e-06,
"loss": 0.1707,
"step": 895
},
{
"epoch": 1.825776872134488,
"grad_norm": 1.083571434020996,
"learning_rate": 2.177358490566038e-06,
"loss": 0.1743,
"step": 896
},
{
"epoch": 1.8278145695364238,
"grad_norm": 1.2324374914169312,
"learning_rate": 2.1735849056603774e-06,
"loss": 0.1876,
"step": 897
},
{
"epoch": 1.8298522669383597,
"grad_norm": 1.1662664413452148,
"learning_rate": 2.1698113207547173e-06,
"loss": 0.172,
"step": 898
},
{
"epoch": 1.8318899643402955,
"grad_norm": 1.0966416597366333,
"learning_rate": 2.166037735849057e-06,
"loss": 0.1784,
"step": 899
},
{
"epoch": 1.8339276617422313,
"grad_norm": 1.0962932109832764,
"learning_rate": 2.1622641509433963e-06,
"loss": 0.1791,
"step": 900
},
{
"epoch": 1.8359653591441671,
"grad_norm": 1.1369909048080444,
"learning_rate": 2.158490566037736e-06,
"loss": 0.1797,
"step": 901
},
{
"epoch": 1.838003056546103,
"grad_norm": 1.1816999912261963,
"learning_rate": 2.1547169811320757e-06,
"loss": 0.169,
"step": 902
},
{
"epoch": 1.8400407539480388,
"grad_norm": 1.1556625366210938,
"learning_rate": 2.150943396226415e-06,
"loss": 0.17,
"step": 903
},
{
"epoch": 1.8420784513499746,
"grad_norm": 1.0892881155014038,
"learning_rate": 2.1471698113207547e-06,
"loss": 0.1652,
"step": 904
},
{
"epoch": 1.8441161487519104,
"grad_norm": 1.3906255960464478,
"learning_rate": 2.1433962264150946e-06,
"loss": 0.1839,
"step": 905
},
{
"epoch": 1.8461538461538463,
"grad_norm": 1.0891425609588623,
"learning_rate": 2.139622641509434e-06,
"loss": 0.1773,
"step": 906
},
{
"epoch": 1.848191543555782,
"grad_norm": 1.1463353633880615,
"learning_rate": 2.1358490566037736e-06,
"loss": 0.1808,
"step": 907
},
{
"epoch": 1.850229240957718,
"grad_norm": 1.086715579032898,
"learning_rate": 2.1320754716981135e-06,
"loss": 0.1804,
"step": 908
},
{
"epoch": 1.8522669383596537,
"grad_norm": 1.102216124534607,
"learning_rate": 2.128301886792453e-06,
"loss": 0.1729,
"step": 909
},
{
"epoch": 1.8543046357615895,
"grad_norm": 1.2313193082809448,
"learning_rate": 2.1245283018867925e-06,
"loss": 0.1906,
"step": 910
},
{
"epoch": 1.8563423331635254,
"grad_norm": 1.3457517623901367,
"learning_rate": 2.120754716981132e-06,
"loss": 0.1675,
"step": 911
},
{
"epoch": 1.8583800305654612,
"grad_norm": 1.1635335683822632,
"learning_rate": 2.116981132075472e-06,
"loss": 0.1775,
"step": 912
},
{
"epoch": 1.8604177279673968,
"grad_norm": 1.2560811042785645,
"learning_rate": 2.1132075471698114e-06,
"loss": 0.1694,
"step": 913
},
{
"epoch": 1.8624554253693326,
"grad_norm": 1.1669859886169434,
"learning_rate": 2.109433962264151e-06,
"loss": 0.1724,
"step": 914
},
{
"epoch": 1.8644931227712684,
"grad_norm": 1.1948050260543823,
"learning_rate": 2.1056603773584908e-06,
"loss": 0.1858,
"step": 915
},
{
"epoch": 1.8665308201732043,
"grad_norm": 1.201643705368042,
"learning_rate": 2.1018867924528303e-06,
"loss": 0.1636,
"step": 916
},
{
"epoch": 1.86856851757514,
"grad_norm": 1.0382106304168701,
"learning_rate": 2.0981132075471698e-06,
"loss": 0.1664,
"step": 917
},
{
"epoch": 1.870606214977076,
"grad_norm": 1.1447466611862183,
"learning_rate": 2.0943396226415097e-06,
"loss": 0.1805,
"step": 918
},
{
"epoch": 1.8726439123790117,
"grad_norm": 1.0567753314971924,
"learning_rate": 2.090566037735849e-06,
"loss": 0.1847,
"step": 919
},
{
"epoch": 1.8746816097809476,
"grad_norm": 1.1714054346084595,
"learning_rate": 2.086792452830189e-06,
"loss": 0.1699,
"step": 920
},
{
"epoch": 1.8767193071828834,
"grad_norm": 1.112230658531189,
"learning_rate": 2.083018867924528e-06,
"loss": 0.1709,
"step": 921
},
{
"epoch": 1.8787570045848192,
"grad_norm": 1.248382329940796,
"learning_rate": 2.079245283018868e-06,
"loss": 0.1854,
"step": 922
},
{
"epoch": 1.8807947019867548,
"grad_norm": 1.0857242345809937,
"learning_rate": 2.075471698113208e-06,
"loss": 0.1647,
"step": 923
},
{
"epoch": 1.8828323993886906,
"grad_norm": 1.1596136093139648,
"learning_rate": 2.0716981132075475e-06,
"loss": 0.1884,
"step": 924
},
{
"epoch": 1.8848700967906264,
"grad_norm": 1.0483487844467163,
"learning_rate": 2.067924528301887e-06,
"loss": 0.1825,
"step": 925
},
{
"epoch": 1.8869077941925623,
"grad_norm": 1.1504698991775513,
"learning_rate": 2.0641509433962264e-06,
"loss": 0.1788,
"step": 926
},
{
"epoch": 1.888945491594498,
"grad_norm": 1.0937446355819702,
"learning_rate": 2.0603773584905664e-06,
"loss": 0.1714,
"step": 927
},
{
"epoch": 1.890983188996434,
"grad_norm": 1.2522544860839844,
"learning_rate": 2.056603773584906e-06,
"loss": 0.1895,
"step": 928
},
{
"epoch": 1.8930208863983697,
"grad_norm": 1.0965933799743652,
"learning_rate": 2.0528301886792453e-06,
"loss": 0.1735,
"step": 929
},
{
"epoch": 1.8950585838003056,
"grad_norm": 1.3030322790145874,
"learning_rate": 2.0490566037735853e-06,
"loss": 0.1705,
"step": 930
},
{
"epoch": 1.8970962812022414,
"grad_norm": 1.1427980661392212,
"learning_rate": 2.0452830188679247e-06,
"loss": 0.1791,
"step": 931
},
{
"epoch": 1.8991339786041772,
"grad_norm": 1.1021360158920288,
"learning_rate": 2.0415094339622642e-06,
"loss": 0.1626,
"step": 932
},
{
"epoch": 1.901171676006113,
"grad_norm": 1.225327968597412,
"learning_rate": 2.037735849056604e-06,
"loss": 0.1746,
"step": 933
},
{
"epoch": 1.9032093734080489,
"grad_norm": 1.4384862184524536,
"learning_rate": 2.0339622641509436e-06,
"loss": 0.1811,
"step": 934
},
{
"epoch": 1.9052470708099847,
"grad_norm": 1.1396024227142334,
"learning_rate": 2.030188679245283e-06,
"loss": 0.1645,
"step": 935
},
{
"epoch": 1.9072847682119205,
"grad_norm": 1.1487840414047241,
"learning_rate": 2.0264150943396226e-06,
"loss": 0.1727,
"step": 936
},
{
"epoch": 1.9093224656138563,
"grad_norm": 1.137575387954712,
"learning_rate": 2.0226415094339625e-06,
"loss": 0.1673,
"step": 937
},
{
"epoch": 1.9113601630157921,
"grad_norm": 1.1038920879364014,
"learning_rate": 2.018867924528302e-06,
"loss": 0.1758,
"step": 938
},
{
"epoch": 1.913397860417728,
"grad_norm": 1.162651777267456,
"learning_rate": 2.0150943396226415e-06,
"loss": 0.1658,
"step": 939
},
{
"epoch": 1.9154355578196638,
"grad_norm": 1.0977519750595093,
"learning_rate": 2.0113207547169814e-06,
"loss": 0.179,
"step": 940
},
{
"epoch": 1.9174732552215996,
"grad_norm": 1.3130261898040771,
"learning_rate": 2.007547169811321e-06,
"loss": 0.1701,
"step": 941
},
{
"epoch": 1.9195109526235354,
"grad_norm": 1.1742639541625977,
"learning_rate": 2.0037735849056604e-06,
"loss": 0.1674,
"step": 942
},
{
"epoch": 1.9215486500254713,
"grad_norm": 1.160561203956604,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.1765,
"step": 943
},
{
"epoch": 1.923586347427407,
"grad_norm": 1.3576925992965698,
"learning_rate": 1.99622641509434e-06,
"loss": 0.1718,
"step": 944
},
{
"epoch": 1.925624044829343,
"grad_norm": 1.101428747177124,
"learning_rate": 1.9924528301886793e-06,
"loss": 0.1843,
"step": 945
},
{
"epoch": 1.9276617422312787,
"grad_norm": 1.271204948425293,
"learning_rate": 1.988679245283019e-06,
"loss": 0.1696,
"step": 946
},
{
"epoch": 1.9296994396332146,
"grad_norm": 1.2334147691726685,
"learning_rate": 1.9849056603773587e-06,
"loss": 0.1713,
"step": 947
},
{
"epoch": 1.9317371370351504,
"grad_norm": 1.2214933633804321,
"learning_rate": 1.981132075471698e-06,
"loss": 0.1819,
"step": 948
},
{
"epoch": 1.9337748344370862,
"grad_norm": 1.1319000720977783,
"learning_rate": 1.9773584905660377e-06,
"loss": 0.1763,
"step": 949
},
{
"epoch": 1.935812531839022,
"grad_norm": 1.0747102499008179,
"learning_rate": 1.9735849056603776e-06,
"loss": 0.179,
"step": 950
},
{
"epoch": 1.9378502292409578,
"grad_norm": 1.1016148328781128,
"learning_rate": 1.969811320754717e-06,
"loss": 0.1723,
"step": 951
},
{
"epoch": 1.9398879266428937,
"grad_norm": 1.3073527812957764,
"learning_rate": 1.9660377358490566e-06,
"loss": 0.18,
"step": 952
},
{
"epoch": 1.9419256240448295,
"grad_norm": 1.2624202966690063,
"learning_rate": 1.9622641509433965e-06,
"loss": 0.1945,
"step": 953
},
{
"epoch": 1.9439633214467653,
"grad_norm": 1.3091782331466675,
"learning_rate": 1.958490566037736e-06,
"loss": 0.1938,
"step": 954
},
{
"epoch": 1.946001018848701,
"grad_norm": 1.136667251586914,
"learning_rate": 1.9547169811320755e-06,
"loss": 0.1808,
"step": 955
},
{
"epoch": 1.9480387162506367,
"grad_norm": 1.1663713455200195,
"learning_rate": 1.950943396226415e-06,
"loss": 0.1735,
"step": 956
},
{
"epoch": 1.9500764136525726,
"grad_norm": 1.2320809364318848,
"learning_rate": 1.947169811320755e-06,
"loss": 0.1757,
"step": 957
},
{
"epoch": 1.9521141110545084,
"grad_norm": 1.174214243888855,
"learning_rate": 1.943396226415095e-06,
"loss": 0.1776,
"step": 958
},
{
"epoch": 1.9541518084564442,
"grad_norm": 1.2423794269561768,
"learning_rate": 1.939622641509434e-06,
"loss": 0.1921,
"step": 959
},
{
"epoch": 1.95618950585838,
"grad_norm": 1.1554875373840332,
"learning_rate": 1.935849056603774e-06,
"loss": 0.1698,
"step": 960
},
{
"epoch": 1.9582272032603159,
"grad_norm": 1.1640571355819702,
"learning_rate": 1.9320754716981133e-06,
"loss": 0.1838,
"step": 961
},
{
"epoch": 1.9602649006622517,
"grad_norm": 1.1926047801971436,
"learning_rate": 1.928301886792453e-06,
"loss": 0.1775,
"step": 962
},
{
"epoch": 1.9623025980641875,
"grad_norm": 1.2760028839111328,
"learning_rate": 1.9245283018867927e-06,
"loss": 0.178,
"step": 963
},
{
"epoch": 1.9643402954661233,
"grad_norm": 1.1897207498550415,
"learning_rate": 1.920754716981132e-06,
"loss": 0.1675,
"step": 964
},
{
"epoch": 1.9663779928680591,
"grad_norm": 1.0810887813568115,
"learning_rate": 1.916981132075472e-06,
"loss": 0.1631,
"step": 965
},
{
"epoch": 1.9684156902699947,
"grad_norm": 1.1327540874481201,
"learning_rate": 1.9132075471698116e-06,
"loss": 0.1834,
"step": 966
},
{
"epoch": 1.9704533876719306,
"grad_norm": 1.0766308307647705,
"learning_rate": 1.909433962264151e-06,
"loss": 0.1756,
"step": 967
},
{
"epoch": 1.9724910850738664,
"grad_norm": 1.1231815814971924,
"learning_rate": 1.9056603773584908e-06,
"loss": 0.1736,
"step": 968
},
{
"epoch": 1.9745287824758022,
"grad_norm": 1.10451078414917,
"learning_rate": 1.9018867924528303e-06,
"loss": 0.1848,
"step": 969
},
{
"epoch": 1.976566479877738,
"grad_norm": 1.114749789237976,
"learning_rate": 1.89811320754717e-06,
"loss": 0.1685,
"step": 970
},
{
"epoch": 1.9786041772796739,
"grad_norm": 1.1218091249465942,
"learning_rate": 1.8943396226415095e-06,
"loss": 0.1644,
"step": 971
},
{
"epoch": 1.9806418746816097,
"grad_norm": 1.1256656646728516,
"learning_rate": 1.8905660377358492e-06,
"loss": 0.1686,
"step": 972
},
{
"epoch": 1.9826795720835455,
"grad_norm": 1.2012169361114502,
"learning_rate": 1.8867924528301889e-06,
"loss": 0.1906,
"step": 973
},
{
"epoch": 1.9847172694854813,
"grad_norm": 1.1859033107757568,
"learning_rate": 1.8830188679245284e-06,
"loss": 0.1818,
"step": 974
},
{
"epoch": 1.9867549668874172,
"grad_norm": 1.1662039756774902,
"learning_rate": 1.879245283018868e-06,
"loss": 0.1928,
"step": 975
},
{
"epoch": 1.988792664289353,
"grad_norm": 1.1107443571090698,
"learning_rate": 1.8754716981132076e-06,
"loss": 0.1701,
"step": 976
},
{
"epoch": 1.9908303616912888,
"grad_norm": 1.1272541284561157,
"learning_rate": 1.8716981132075473e-06,
"loss": 0.1676,
"step": 977
},
{
"epoch": 1.9928680590932246,
"grad_norm": 1.158721923828125,
"learning_rate": 1.8679245283018868e-06,
"loss": 0.1642,
"step": 978
},
{
"epoch": 1.9949057564951604,
"grad_norm": 1.1401432752609253,
"learning_rate": 1.8641509433962265e-06,
"loss": 0.1674,
"step": 979
},
{
"epoch": 1.9969434538970963,
"grad_norm": 1.1835023164749146,
"learning_rate": 1.8603773584905664e-06,
"loss": 0.1812,
"step": 980
},
{
"epoch": 1.998981151299032,
"grad_norm": 1.2545579671859741,
"learning_rate": 1.8566037735849056e-06,
"loss": 0.1858,
"step": 981
},
{
"epoch": 2.001018848700968,
"grad_norm": 1.0314708948135376,
"learning_rate": 1.8528301886792456e-06,
"loss": 0.1435,
"step": 982
}
],
"logging_steps": 1.0,
"max_steps": 1473,
"num_input_tokens_seen": 0,
"num_train_epochs": 4,
"save_steps": 491,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.3189159645267624e+19,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}