nttx's picture
Training in progress, step 100, checkpoint
aadf04f verified
raw
history blame
18.8 kB
{
"best_metric": 0.5659072399139404,
"best_model_checkpoint": "miner_id_24/checkpoint-100",
"epoch": 0.1502065339842283,
"eval_steps": 50,
"global_step": 100,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0015020653398422831,
"grad_norm": 0.48013806343078613,
"learning_rate": 1e-05,
"loss": 1.1844,
"step": 1
},
{
"epoch": 0.0015020653398422831,
"eval_loss": 1.4352283477783203,
"eval_runtime": 33.3798,
"eval_samples_per_second": 33.613,
"eval_steps_per_second": 8.418,
"step": 1
},
{
"epoch": 0.0030041306796845663,
"grad_norm": 0.42095404863357544,
"learning_rate": 2e-05,
"loss": 1.2759,
"step": 2
},
{
"epoch": 0.004506196019526849,
"grad_norm": 0.40659624338150024,
"learning_rate": 3e-05,
"loss": 1.3025,
"step": 3
},
{
"epoch": 0.0060082613593691325,
"grad_norm": 0.41802921891212463,
"learning_rate": 4e-05,
"loss": 1.3053,
"step": 4
},
{
"epoch": 0.007510326699211416,
"grad_norm": 0.4731360673904419,
"learning_rate": 5e-05,
"loss": 1.305,
"step": 5
},
{
"epoch": 0.009012392039053699,
"grad_norm": 0.4001414179801941,
"learning_rate": 6e-05,
"loss": 1.301,
"step": 6
},
{
"epoch": 0.010514457378895982,
"grad_norm": 0.3494016230106354,
"learning_rate": 7e-05,
"loss": 1.2796,
"step": 7
},
{
"epoch": 0.012016522718738265,
"grad_norm": 0.25853636860847473,
"learning_rate": 8e-05,
"loss": 1.2454,
"step": 8
},
{
"epoch": 0.013518588058580548,
"grad_norm": 0.36195915937423706,
"learning_rate": 9e-05,
"loss": 1.2594,
"step": 9
},
{
"epoch": 0.015020653398422831,
"grad_norm": 0.41546186804771423,
"learning_rate": 0.0001,
"loss": 1.2273,
"step": 10
},
{
"epoch": 0.016522718738265114,
"grad_norm": 0.38150355219841003,
"learning_rate": 9.999316524962345e-05,
"loss": 1.1949,
"step": 11
},
{
"epoch": 0.018024784078107398,
"grad_norm": 0.3044074773788452,
"learning_rate": 9.997266286704631e-05,
"loss": 1.1619,
"step": 12
},
{
"epoch": 0.01952684941794968,
"grad_norm": 0.23984934389591217,
"learning_rate": 9.993849845741524e-05,
"loss": 1.1339,
"step": 13
},
{
"epoch": 0.021028914757791964,
"grad_norm": 0.24785546958446503,
"learning_rate": 9.989068136093873e-05,
"loss": 1.0967,
"step": 14
},
{
"epoch": 0.022530980097634247,
"grad_norm": 0.25208961963653564,
"learning_rate": 9.98292246503335e-05,
"loss": 1.074,
"step": 15
},
{
"epoch": 0.02403304543747653,
"grad_norm": 0.26858824491500854,
"learning_rate": 9.975414512725057e-05,
"loss": 1.0323,
"step": 16
},
{
"epoch": 0.025535110777318813,
"grad_norm": 0.2699333131313324,
"learning_rate": 9.966546331768191e-05,
"loss": 1.0193,
"step": 17
},
{
"epoch": 0.027037176117161096,
"grad_norm": 0.29868969321250916,
"learning_rate": 9.956320346634876e-05,
"loss": 0.9937,
"step": 18
},
{
"epoch": 0.02853924145700338,
"grad_norm": 0.2335326373577118,
"learning_rate": 9.944739353007344e-05,
"loss": 0.9666,
"step": 19
},
{
"epoch": 0.030041306796845663,
"grad_norm": 0.25925031304359436,
"learning_rate": 9.931806517013612e-05,
"loss": 0.9465,
"step": 20
},
{
"epoch": 0.031543372136687946,
"grad_norm": 0.3260612189769745,
"learning_rate": 9.917525374361912e-05,
"loss": 0.9378,
"step": 21
},
{
"epoch": 0.03304543747653023,
"grad_norm": 0.2550596594810486,
"learning_rate": 9.901899829374047e-05,
"loss": 0.9038,
"step": 22
},
{
"epoch": 0.03454750281637251,
"grad_norm": 0.28878340125083923,
"learning_rate": 9.884934153917997e-05,
"loss": 0.8962,
"step": 23
},
{
"epoch": 0.036049568156214795,
"grad_norm": 0.27984026074409485,
"learning_rate": 9.86663298624003e-05,
"loss": 0.8538,
"step": 24
},
{
"epoch": 0.03755163349605708,
"grad_norm": 0.2650708556175232,
"learning_rate": 9.847001329696653e-05,
"loss": 0.8394,
"step": 25
},
{
"epoch": 0.03905369883589936,
"grad_norm": 0.24351182579994202,
"learning_rate": 9.826044551386744e-05,
"loss": 0.8104,
"step": 26
},
{
"epoch": 0.040555764175741645,
"grad_norm": 0.3326224088668823,
"learning_rate": 9.803768380684242e-05,
"loss": 0.794,
"step": 27
},
{
"epoch": 0.04205782951558393,
"grad_norm": 0.298579603433609,
"learning_rate": 9.780178907671789e-05,
"loss": 0.7783,
"step": 28
},
{
"epoch": 0.04355989485542621,
"grad_norm": 0.2561696469783783,
"learning_rate": 9.755282581475769e-05,
"loss": 0.7657,
"step": 29
},
{
"epoch": 0.045061960195268494,
"grad_norm": 0.2900291383266449,
"learning_rate": 9.729086208503174e-05,
"loss": 0.7382,
"step": 30
},
{
"epoch": 0.04656402553511078,
"grad_norm": 0.2575244605541229,
"learning_rate": 9.701596950580806e-05,
"loss": 0.723,
"step": 31
},
{
"epoch": 0.04806609087495306,
"grad_norm": 0.34842997789382935,
"learning_rate": 9.672822322997305e-05,
"loss": 0.7148,
"step": 32
},
{
"epoch": 0.04956815621479534,
"grad_norm": 0.2418001890182495,
"learning_rate": 9.642770192448536e-05,
"loss": 0.6728,
"step": 33
},
{
"epoch": 0.051070221554637626,
"grad_norm": 0.37016379833221436,
"learning_rate": 9.611448774886924e-05,
"loss": 0.6626,
"step": 34
},
{
"epoch": 0.05257228689447991,
"grad_norm": 0.3161463439464569,
"learning_rate": 9.578866633275288e-05,
"loss": 0.6567,
"step": 35
},
{
"epoch": 0.05407435223432219,
"grad_norm": 0.4346078038215637,
"learning_rate": 9.545032675245813e-05,
"loss": 0.6694,
"step": 36
},
{
"epoch": 0.055576417574164476,
"grad_norm": 0.435017853975296,
"learning_rate": 9.509956150664796e-05,
"loss": 0.6299,
"step": 37
},
{
"epoch": 0.05707848291400676,
"grad_norm": 0.3490477204322815,
"learning_rate": 9.473646649103818e-05,
"loss": 0.6261,
"step": 38
},
{
"epoch": 0.05858054825384904,
"grad_norm": 0.5194304585456848,
"learning_rate": 9.43611409721806e-05,
"loss": 0.5896,
"step": 39
},
{
"epoch": 0.060082613593691325,
"grad_norm": 0.2708154618740082,
"learning_rate": 9.397368756032445e-05,
"loss": 0.6089,
"step": 40
},
{
"epoch": 0.06158467893353361,
"grad_norm": 0.4544765055179596,
"learning_rate": 9.357421218136386e-05,
"loss": 0.5747,
"step": 41
},
{
"epoch": 0.06308674427337589,
"grad_norm": 0.3197058439254761,
"learning_rate": 9.316282404787871e-05,
"loss": 0.5633,
"step": 42
},
{
"epoch": 0.06458880961321818,
"grad_norm": 0.49233025312423706,
"learning_rate": 9.273963562927695e-05,
"loss": 0.5625,
"step": 43
},
{
"epoch": 0.06609087495306046,
"grad_norm": 0.31287699937820435,
"learning_rate": 9.230476262104677e-05,
"loss": 0.5604,
"step": 44
},
{
"epoch": 0.06759294029290275,
"grad_norm": 0.4544554054737091,
"learning_rate": 9.185832391312644e-05,
"loss": 0.5784,
"step": 45
},
{
"epoch": 0.06909500563274502,
"grad_norm": 0.4347470998764038,
"learning_rate": 9.140044155740101e-05,
"loss": 0.5394,
"step": 46
},
{
"epoch": 0.07059707097258731,
"grad_norm": 0.4225442707538605,
"learning_rate": 9.093124073433463e-05,
"loss": 0.55,
"step": 47
},
{
"epoch": 0.07209913631242959,
"grad_norm": 0.4641411602497101,
"learning_rate": 9.045084971874738e-05,
"loss": 0.5168,
"step": 48
},
{
"epoch": 0.07360120165227188,
"grad_norm": 0.5490090847015381,
"learning_rate": 8.995939984474624e-05,
"loss": 0.545,
"step": 49
},
{
"epoch": 0.07510326699211416,
"grad_norm": 0.5186936855316162,
"learning_rate": 8.945702546981969e-05,
"loss": 0.4797,
"step": 50
},
{
"epoch": 0.07510326699211416,
"eval_loss": 0.9435171484947205,
"eval_runtime": 33.2926,
"eval_samples_per_second": 33.701,
"eval_steps_per_second": 8.44,
"step": 50
},
{
"epoch": 0.07660533233195645,
"grad_norm": 5.346027851104736,
"learning_rate": 8.894386393810563e-05,
"loss": 1.3858,
"step": 51
},
{
"epoch": 0.07810739767179872,
"grad_norm": 4.529382228851318,
"learning_rate": 8.842005554284296e-05,
"loss": 1.1019,
"step": 52
},
{
"epoch": 0.07960946301164101,
"grad_norm": 2.0946102142333984,
"learning_rate": 8.788574348801675e-05,
"loss": 0.7455,
"step": 53
},
{
"epoch": 0.08111152835148329,
"grad_norm": 0.4784291386604309,
"learning_rate": 8.73410738492077e-05,
"loss": 0.6575,
"step": 54
},
{
"epoch": 0.08261359369132558,
"grad_norm": 0.43049320578575134,
"learning_rate": 8.678619553365659e-05,
"loss": 0.6418,
"step": 55
},
{
"epoch": 0.08411565903116786,
"grad_norm": 0.3956131637096405,
"learning_rate": 8.622126023955446e-05,
"loss": 0.6338,
"step": 56
},
{
"epoch": 0.08561772437101015,
"grad_norm": 0.3648943603038788,
"learning_rate": 8.564642241456986e-05,
"loss": 0.6281,
"step": 57
},
{
"epoch": 0.08711978971085242,
"grad_norm": 0.265221506357193,
"learning_rate": 8.506183921362443e-05,
"loss": 0.6154,
"step": 58
},
{
"epoch": 0.08862185505069471,
"grad_norm": 0.3286784887313843,
"learning_rate": 8.44676704559283e-05,
"loss": 0.6275,
"step": 59
},
{
"epoch": 0.09012392039053699,
"grad_norm": 0.36223769187927246,
"learning_rate": 8.386407858128706e-05,
"loss": 0.6223,
"step": 60
},
{
"epoch": 0.09162598573037928,
"grad_norm": 0.2875317931175232,
"learning_rate": 8.32512286056924e-05,
"loss": 0.614,
"step": 61
},
{
"epoch": 0.09312805107022155,
"grad_norm": 0.24351203441619873,
"learning_rate": 8.262928807620843e-05,
"loss": 0.5952,
"step": 62
},
{
"epoch": 0.09463011641006384,
"grad_norm": 0.2812121510505676,
"learning_rate": 8.199842702516583e-05,
"loss": 0.6036,
"step": 63
},
{
"epoch": 0.09613218174990612,
"grad_norm": 0.2832166850566864,
"learning_rate": 8.135881792367686e-05,
"loss": 0.5976,
"step": 64
},
{
"epoch": 0.09763424708974841,
"grad_norm": 0.2672559916973114,
"learning_rate": 8.07106356344834e-05,
"loss": 0.5743,
"step": 65
},
{
"epoch": 0.09913631242959069,
"grad_norm": 0.25090304017066956,
"learning_rate": 8.005405736415126e-05,
"loss": 0.5795,
"step": 66
},
{
"epoch": 0.10063837776943298,
"grad_norm": 0.3040817677974701,
"learning_rate": 7.938926261462366e-05,
"loss": 0.5613,
"step": 67
},
{
"epoch": 0.10214044310927525,
"grad_norm": 0.23814338445663452,
"learning_rate": 7.871643313414718e-05,
"loss": 0.5521,
"step": 68
},
{
"epoch": 0.10364250844911754,
"grad_norm": 0.2542641758918762,
"learning_rate": 7.803575286758364e-05,
"loss": 0.5535,
"step": 69
},
{
"epoch": 0.10514457378895982,
"grad_norm": 0.2717762887477875,
"learning_rate": 7.734740790612136e-05,
"loss": 0.5361,
"step": 70
},
{
"epoch": 0.10664663912880211,
"grad_norm": 0.2743045389652252,
"learning_rate": 7.66515864363997e-05,
"loss": 0.5338,
"step": 71
},
{
"epoch": 0.10814870446864439,
"grad_norm": 0.22834768891334534,
"learning_rate": 7.594847868906076e-05,
"loss": 0.5238,
"step": 72
},
{
"epoch": 0.10965076980848668,
"grad_norm": 0.3098483085632324,
"learning_rate": 7.52382768867422e-05,
"loss": 0.5278,
"step": 73
},
{
"epoch": 0.11115283514832895,
"grad_norm": 0.2830328941345215,
"learning_rate": 7.452117519152542e-05,
"loss": 0.5089,
"step": 74
},
{
"epoch": 0.11265490048817124,
"grad_norm": 0.2570495009422302,
"learning_rate": 7.379736965185368e-05,
"loss": 0.4981,
"step": 75
},
{
"epoch": 0.11415696582801352,
"grad_norm": 0.27931836247444153,
"learning_rate": 7.30670581489344e-05,
"loss": 0.4863,
"step": 76
},
{
"epoch": 0.11565903116785581,
"grad_norm": 0.2561793625354767,
"learning_rate": 7.233044034264034e-05,
"loss": 0.4959,
"step": 77
},
{
"epoch": 0.11716109650769808,
"grad_norm": 0.29976847767829895,
"learning_rate": 7.158771761692464e-05,
"loss": 0.4882,
"step": 78
},
{
"epoch": 0.11866316184754037,
"grad_norm": 0.27593934535980225,
"learning_rate": 7.083909302476453e-05,
"loss": 0.4959,
"step": 79
},
{
"epoch": 0.12016522718738265,
"grad_norm": 0.2488517165184021,
"learning_rate": 7.008477123264848e-05,
"loss": 0.4699,
"step": 80
},
{
"epoch": 0.12166729252722494,
"grad_norm": 0.19852301478385925,
"learning_rate": 6.932495846462261e-05,
"loss": 0.4576,
"step": 81
},
{
"epoch": 0.12316935786706722,
"grad_norm": 0.2308630794286728,
"learning_rate": 6.855986244591104e-05,
"loss": 0.4519,
"step": 82
},
{
"epoch": 0.1246714232069095,
"grad_norm": 0.28317439556121826,
"learning_rate": 6.778969234612584e-05,
"loss": 0.4773,
"step": 83
},
{
"epoch": 0.12617348854675178,
"grad_norm": 0.280638188123703,
"learning_rate": 6.701465872208216e-05,
"loss": 0.4727,
"step": 84
},
{
"epoch": 0.12767555388659407,
"grad_norm": 0.29901838302612305,
"learning_rate": 6.623497346023418e-05,
"loss": 0.463,
"step": 85
},
{
"epoch": 0.12917761922643636,
"grad_norm": 0.32130149006843567,
"learning_rate": 6.545084971874738e-05,
"loss": 0.4546,
"step": 86
},
{
"epoch": 0.13067968456627863,
"grad_norm": 0.24742907285690308,
"learning_rate": 6.466250186922325e-05,
"loss": 0.4519,
"step": 87
},
{
"epoch": 0.13218174990612092,
"grad_norm": 0.2867355942726135,
"learning_rate": 6.387014543809223e-05,
"loss": 0.4445,
"step": 88
},
{
"epoch": 0.1336838152459632,
"grad_norm": 0.27279844880104065,
"learning_rate": 6.307399704769099e-05,
"loss": 0.4328,
"step": 89
},
{
"epoch": 0.1351858805858055,
"grad_norm": 0.25699475407600403,
"learning_rate": 6.227427435703997e-05,
"loss": 0.4397,
"step": 90
},
{
"epoch": 0.13668794592564776,
"grad_norm": 0.2791455090045929,
"learning_rate": 6.147119600233758e-05,
"loss": 0.4523,
"step": 91
},
{
"epoch": 0.13819001126549005,
"grad_norm": 0.3582814335823059,
"learning_rate": 6.066498153718735e-05,
"loss": 0.4553,
"step": 92
},
{
"epoch": 0.13969207660533234,
"grad_norm": 0.28347545862197876,
"learning_rate": 5.985585137257401e-05,
"loss": 0.4314,
"step": 93
},
{
"epoch": 0.14119414194517463,
"grad_norm": 0.2645236551761627,
"learning_rate": 5.90440267166055e-05,
"loss": 0.4493,
"step": 94
},
{
"epoch": 0.1426962072850169,
"grad_norm": 0.308789998292923,
"learning_rate": 5.8229729514036705e-05,
"loss": 0.4487,
"step": 95
},
{
"epoch": 0.14419827262485918,
"grad_norm": 0.27754440903663635,
"learning_rate": 5.74131823855921e-05,
"loss": 0.4236,
"step": 96
},
{
"epoch": 0.14570033796470147,
"grad_norm": 0.29128730297088623,
"learning_rate": 5.6594608567103456e-05,
"loss": 0.4523,
"step": 97
},
{
"epoch": 0.14720240330454376,
"grad_norm": 0.4283706247806549,
"learning_rate": 5.577423184847932e-05,
"loss": 0.4624,
"step": 98
},
{
"epoch": 0.14870446864438602,
"grad_norm": 0.3860466480255127,
"learning_rate": 5.495227651252315e-05,
"loss": 0.4384,
"step": 99
},
{
"epoch": 0.1502065339842283,
"grad_norm": 0.4692773222923279,
"learning_rate": 5.4128967273616625e-05,
"loss": 0.4551,
"step": 100
},
{
"epoch": 0.1502065339842283,
"eval_loss": 0.5659072399139404,
"eval_runtime": 33.2992,
"eval_samples_per_second": 33.694,
"eval_steps_per_second": 8.439,
"step": 100
}
],
"logging_steps": 1,
"max_steps": 200,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"EarlyStoppingCallback": {
"args": {
"early_stopping_patience": 5,
"early_stopping_threshold": 0.0
},
"attributes": {
"early_stopping_patience_counter": 0
}
},
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 3.3173127051608064e+16,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}