Pravincoder's picture
Upload folder using huggingface_hub
f152b82
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 399.0662251655629,
"eval_steps": 500,
"global_step": 4000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 2.07,
"learning_rate": 2.4834437086092715e-05,
"loss": 3.036,
"step": 30
},
{
"epoch": 5.07,
"learning_rate": 4.966887417218543e-05,
"loss": 2.727,
"step": 60
},
{
"epoch": 8.07,
"learning_rate": 7.450331125827815e-05,
"loss": 2.4733,
"step": 90
},
{
"epoch": 11.07,
"learning_rate": 9.933774834437086e-05,
"loss": 2.3326,
"step": 120
},
{
"epoch": 14.07,
"learning_rate": 0.00012417218543046358,
"loss": 2.2252,
"step": 150
},
{
"epoch": 17.07,
"learning_rate": 0.0001490066225165563,
"loss": 2.1346,
"step": 180
},
{
"epoch": 20.07,
"learning_rate": 0.000173841059602649,
"loss": 2.0513,
"step": 210
},
{
"epoch": 23.07,
"learning_rate": 0.00019867549668874172,
"loss": 1.9759,
"step": 240
},
{
"epoch": 26.07,
"learning_rate": 0.00022350993377483444,
"loss": 1.9071,
"step": 270
},
{
"epoch": 29.07,
"learning_rate": 0.00024834437086092715,
"loss": 1.838,
"step": 300
},
{
"epoch": 32.07,
"learning_rate": 0.0002731788079470199,
"loss": 1.7631,
"step": 330
},
{
"epoch": 35.07,
"learning_rate": 0.0002980132450331126,
"loss": 1.6968,
"step": 360
},
{
"epoch": 38.07,
"learning_rate": 0.00032284768211920527,
"loss": 1.6119,
"step": 390
},
{
"epoch": 41.07,
"learning_rate": 0.000347682119205298,
"loss": 1.5422,
"step": 420
},
{
"epoch": 44.07,
"learning_rate": 0.00037251655629139076,
"loss": 1.4734,
"step": 450
},
{
"epoch": 47.07,
"learning_rate": 0.00039735099337748344,
"loss": 1.3984,
"step": 480
},
{
"epoch": 50.07,
"learning_rate": 0.0004221854304635762,
"loss": 1.3118,
"step": 510
},
{
"epoch": 53.07,
"learning_rate": 0.0004470198675496689,
"loss": 1.2615,
"step": 540
},
{
"epoch": 56.07,
"learning_rate": 0.00047185430463576156,
"loss": 1.1864,
"step": 570
},
{
"epoch": 59.07,
"learning_rate": 0.0004966887417218543,
"loss": 1.1384,
"step": 600
},
{
"epoch": 62.07,
"learning_rate": 0.000521523178807947,
"loss": 1.0538,
"step": 630
},
{
"epoch": 65.07,
"learning_rate": 0.0005463576158940398,
"loss": 1.0015,
"step": 660
},
{
"epoch": 68.07,
"learning_rate": 0.0005711920529801324,
"loss": 0.939,
"step": 690
},
{
"epoch": 71.07,
"learning_rate": 0.0005960264900662252,
"loss": 0.8865,
"step": 720
},
{
"epoch": 74.07,
"learning_rate": 0.0006208609271523179,
"loss": 0.8369,
"step": 750
},
{
"epoch": 77.07,
"learning_rate": 0.0006456953642384105,
"loss": 0.7761,
"step": 780
},
{
"epoch": 80.07,
"learning_rate": 0.0006705298013245033,
"loss": 0.7468,
"step": 810
},
{
"epoch": 83.07,
"learning_rate": 0.000695364238410596,
"loss": 0.6898,
"step": 840
},
{
"epoch": 86.07,
"learning_rate": 0.0007201986754966887,
"loss": 0.6711,
"step": 870
},
{
"epoch": 89.07,
"learning_rate": 0.0007450331125827815,
"loss": 0.6476,
"step": 900
},
{
"epoch": 92.07,
"learning_rate": 0.0007698675496688743,
"loss": 0.6337,
"step": 930
},
{
"epoch": 95.07,
"learning_rate": 0.0007947019867549669,
"loss": 0.5984,
"step": 960
},
{
"epoch": 98.07,
"learning_rate": 0.0008195364238410596,
"loss": 0.5906,
"step": 990
},
{
"epoch": 101.07,
"learning_rate": 0.0008443708609271524,
"loss": 0.5713,
"step": 1020
},
{
"epoch": 104.07,
"learning_rate": 0.000869205298013245,
"loss": 0.5323,
"step": 1050
},
{
"epoch": 107.07,
"learning_rate": 0.0008940397350993377,
"loss": 0.5379,
"step": 1080
},
{
"epoch": 110.07,
"learning_rate": 0.0009188741721854305,
"loss": 0.5165,
"step": 1110
},
{
"epoch": 113.07,
"learning_rate": 0.0009437086092715231,
"loss": 0.5117,
"step": 1140
},
{
"epoch": 116.07,
"learning_rate": 0.0009685430463576159,
"loss": 0.4923,
"step": 1170
},
{
"epoch": 119.07,
"learning_rate": 0.0009933774834437086,
"loss": 0.4754,
"step": 1200
},
{
"epoch": 122.07,
"learning_rate": 0.0010182119205298015,
"loss": 0.4723,
"step": 1230
},
{
"epoch": 125.07,
"learning_rate": 0.001043046357615894,
"loss": 0.4699,
"step": 1260
},
{
"epoch": 128.07,
"learning_rate": 0.0010678807947019867,
"loss": 0.4432,
"step": 1290
},
{
"epoch": 131.07,
"learning_rate": 0.0010927152317880796,
"loss": 0.4537,
"step": 1320
},
{
"epoch": 134.07,
"learning_rate": 0.0011175496688741722,
"loss": 0.4658,
"step": 1350
},
{
"epoch": 137.07,
"learning_rate": 0.0011423841059602648,
"loss": 0.4134,
"step": 1380
},
{
"epoch": 140.07,
"learning_rate": 0.0011672185430463577,
"loss": 0.4326,
"step": 1410
},
{
"epoch": 143.07,
"learning_rate": 0.0011920529801324503,
"loss": 0.452,
"step": 1440
},
{
"epoch": 146.07,
"learning_rate": 0.001216887417218543,
"loss": 0.4637,
"step": 1470
},
{
"epoch": 149.07,
"learning_rate": 0.0012417218543046358,
"loss": 0.4336,
"step": 1500
},
{
"epoch": 152.07,
"learning_rate": 0.0012665562913907287,
"loss": 0.4468,
"step": 1530
},
{
"epoch": 155.07,
"learning_rate": 0.001291390728476821,
"loss": 0.4415,
"step": 1560
},
{
"epoch": 158.07,
"learning_rate": 0.001316225165562914,
"loss": 0.4216,
"step": 1590
},
{
"epoch": 161.07,
"learning_rate": 0.0013410596026490066,
"loss": 0.4268,
"step": 1620
},
{
"epoch": 164.07,
"learning_rate": 0.0013658940397350994,
"loss": 0.4177,
"step": 1650
},
{
"epoch": 167.07,
"learning_rate": 0.001390728476821192,
"loss": 0.4439,
"step": 1680
},
{
"epoch": 170.07,
"learning_rate": 0.001415562913907285,
"loss": 0.444,
"step": 1710
},
{
"epoch": 173.07,
"learning_rate": 0.0014403973509933773,
"loss": 0.4364,
"step": 1740
},
{
"epoch": 176.07,
"learning_rate": 0.0014652317880794702,
"loss": 0.4292,
"step": 1770
},
{
"epoch": 179.07,
"learning_rate": 0.001490066225165563,
"loss": 0.4343,
"step": 1800
},
{
"epoch": 182.07,
"learning_rate": 0.0015149006622516557,
"loss": 0.4513,
"step": 1830
},
{
"epoch": 185.07,
"learning_rate": 0.0015397350993377485,
"loss": 0.453,
"step": 1860
},
{
"epoch": 188.07,
"learning_rate": 0.0015645695364238411,
"loss": 0.4163,
"step": 1890
},
{
"epoch": 191.07,
"learning_rate": 0.0015894039735099338,
"loss": 0.441,
"step": 1920
},
{
"epoch": 194.07,
"learning_rate": 0.0016142384105960264,
"loss": 0.4564,
"step": 1950
},
{
"epoch": 197.07,
"learning_rate": 0.0016390728476821193,
"loss": 0.4733,
"step": 1980
},
{
"epoch": 200.07,
"learning_rate": 0.001663907284768212,
"loss": 0.4634,
"step": 2010
},
{
"epoch": 203.07,
"learning_rate": 0.0016887417218543047,
"loss": 0.4723,
"step": 2040
},
{
"epoch": 206.07,
"learning_rate": 0.0017135761589403972,
"loss": 0.4666,
"step": 2070
},
{
"epoch": 209.07,
"learning_rate": 0.00173841059602649,
"loss": 0.4924,
"step": 2100
},
{
"epoch": 212.07,
"learning_rate": 0.0017632450331125829,
"loss": 0.4821,
"step": 2130
},
{
"epoch": 215.07,
"learning_rate": 0.0017880794701986755,
"loss": 0.48,
"step": 2160
},
{
"epoch": 218.07,
"learning_rate": 0.0018129139072847683,
"loss": 0.5119,
"step": 2190
},
{
"epoch": 221.07,
"learning_rate": 0.001837748344370861,
"loss": 0.5396,
"step": 2220
},
{
"epoch": 224.07,
"learning_rate": 0.0018625827814569538,
"loss": 0.5597,
"step": 2250
},
{
"epoch": 227.07,
"learning_rate": 0.0018874172185430462,
"loss": 0.5535,
"step": 2280
},
{
"epoch": 230.07,
"learning_rate": 0.001912251655629139,
"loss": 0.5951,
"step": 2310
},
{
"epoch": 233.07,
"learning_rate": 0.0019370860927152317,
"loss": 0.553,
"step": 2340
},
{
"epoch": 236.07,
"learning_rate": 0.0019619205298013246,
"loss": 0.5303,
"step": 2370
},
{
"epoch": 239.07,
"learning_rate": 0.001986754966887417,
"loss": 0.5958,
"step": 2400
},
{
"epoch": 242.07,
"learning_rate": 0.00201158940397351,
"loss": 0.6002,
"step": 2430
},
{
"epoch": 245.07,
"learning_rate": 0.002036423841059603,
"loss": 0.6858,
"step": 2460
},
{
"epoch": 248.07,
"learning_rate": 0.0020612582781456956,
"loss": 0.6152,
"step": 2490
},
{
"epoch": 251.07,
"learning_rate": 0.002086092715231788,
"loss": 0.6171,
"step": 2520
},
{
"epoch": 254.07,
"learning_rate": 0.002110927152317881,
"loss": 0.6383,
"step": 2550
},
{
"epoch": 257.07,
"learning_rate": 0.0021357615894039735,
"loss": 0.7459,
"step": 2580
},
{
"epoch": 260.07,
"learning_rate": 0.002160596026490066,
"loss": 0.7926,
"step": 2610
},
{
"epoch": 263.07,
"learning_rate": 0.002185430463576159,
"loss": 0.8945,
"step": 2640
},
{
"epoch": 266.07,
"learning_rate": 0.002210264900662252,
"loss": 0.8908,
"step": 2670
},
{
"epoch": 269.07,
"learning_rate": 0.0022350993377483444,
"loss": 0.9268,
"step": 2700
},
{
"epoch": 272.07,
"learning_rate": 0.002259933774834437,
"loss": 0.8545,
"step": 2730
},
{
"epoch": 275.07,
"learning_rate": 0.0022847682119205297,
"loss": 0.8369,
"step": 2760
},
{
"epoch": 278.07,
"learning_rate": 0.0023096026490066228,
"loss": 0.8851,
"step": 2790
},
{
"epoch": 281.07,
"learning_rate": 0.0023344370860927154,
"loss": 0.8496,
"step": 2820
},
{
"epoch": 284.07,
"learning_rate": 0.002359271523178808,
"loss": 0.8121,
"step": 2850
},
{
"epoch": 287.07,
"learning_rate": 0.0023841059602649007,
"loss": 0.76,
"step": 2880
},
{
"epoch": 290.07,
"learning_rate": 0.0024089403973509937,
"loss": 0.8552,
"step": 2910
},
{
"epoch": 293.07,
"learning_rate": 0.002433774834437086,
"loss": 0.9308,
"step": 2940
},
{
"epoch": 296.07,
"learning_rate": 0.002458609271523179,
"loss": 0.9471,
"step": 2970
},
{
"epoch": 299.07,
"learning_rate": 0.0024834437086092716,
"loss": 0.9702,
"step": 3000
},
{
"epoch": 302.07,
"learning_rate": 0.0025082781456953643,
"loss": 1.0656,
"step": 3030
},
{
"epoch": 305.07,
"learning_rate": 0.0025331125827814573,
"loss": 1.0608,
"step": 3060
},
{
"epoch": 308.07,
"learning_rate": 0.0025579470198675495,
"loss": 1.0397,
"step": 3090
},
{
"epoch": 311.07,
"learning_rate": 0.002582781456953642,
"loss": 1.0805,
"step": 3120
},
{
"epoch": 314.07,
"learning_rate": 0.0026076158940397352,
"loss": 1.1374,
"step": 3150
},
{
"epoch": 317.07,
"learning_rate": 0.002632450331125828,
"loss": 1.4001,
"step": 3180
},
{
"epoch": 320.07,
"learning_rate": 0.0026572847682119205,
"loss": 1.5648,
"step": 3210
},
{
"epoch": 323.07,
"learning_rate": 0.002682119205298013,
"loss": 1.7286,
"step": 3240
},
{
"epoch": 326.07,
"learning_rate": 0.002706953642384106,
"loss": 1.5511,
"step": 3270
},
{
"epoch": 329.07,
"learning_rate": 0.002731788079470199,
"loss": 1.3367,
"step": 3300
},
{
"epoch": 332.07,
"learning_rate": 0.0027566225165562915,
"loss": 1.2894,
"step": 3330
},
{
"epoch": 335.07,
"learning_rate": 0.002781456953642384,
"loss": 1.3288,
"step": 3360
},
{
"epoch": 338.07,
"learning_rate": 0.002806291390728477,
"loss": 1.2965,
"step": 3390
},
{
"epoch": 341.07,
"learning_rate": 0.00283112582781457,
"loss": 1.1965,
"step": 3420
},
{
"epoch": 344.07,
"learning_rate": 0.002855960264900662,
"loss": 1.1625,
"step": 3450
},
{
"epoch": 347.07,
"learning_rate": 0.0028807947019867546,
"loss": 1.2772,
"step": 3480
},
{
"epoch": 350.07,
"learning_rate": 0.0029056291390728477,
"loss": 1.2967,
"step": 3510
},
{
"epoch": 353.07,
"learning_rate": 0.0029304635761589403,
"loss": 1.5922,
"step": 3540
},
{
"epoch": 356.07,
"learning_rate": 0.002955298013245033,
"loss": 1.5614,
"step": 3570
},
{
"epoch": 359.07,
"learning_rate": 0.002980132450331126,
"loss": 1.9061,
"step": 3600
},
{
"epoch": 362.07,
"learning_rate": 0.0030049668874172187,
"loss": 2.0664,
"step": 3630
},
{
"epoch": 365.07,
"learning_rate": 0.0030298013245033113,
"loss": 1.8159,
"step": 3660
},
{
"epoch": 368.07,
"learning_rate": 0.003054635761589404,
"loss": 1.6505,
"step": 3690
},
{
"epoch": 371.07,
"learning_rate": 0.003079470198675497,
"loss": 1.4858,
"step": 3720
},
{
"epoch": 374.07,
"learning_rate": 0.0031043046357615896,
"loss": 1.4008,
"step": 3750
},
{
"epoch": 377.07,
"learning_rate": 0.0031291390728476823,
"loss": 1.2804,
"step": 3780
},
{
"epoch": 380.07,
"learning_rate": 0.0031539735099337745,
"loss": 1.3041,
"step": 3810
},
{
"epoch": 383.07,
"learning_rate": 0.0031788079470198675,
"loss": 1.3115,
"step": 3840
},
{
"epoch": 386.07,
"learning_rate": 0.00320364238410596,
"loss": 1.3194,
"step": 3870
},
{
"epoch": 389.07,
"learning_rate": 0.003228476821192053,
"loss": 1.3681,
"step": 3900
},
{
"epoch": 392.07,
"learning_rate": 0.003253311258278146,
"loss": 1.3676,
"step": 3930
},
{
"epoch": 395.07,
"learning_rate": 0.0032781456953642385,
"loss": 1.702,
"step": 3960
},
{
"epoch": 398.07,
"learning_rate": 0.003302980132450331,
"loss": 2.7594,
"step": 3990
}
],
"logging_steps": 30,
"max_steps": 60400,
"num_train_epochs": 400,
"save_steps": 500,
"total_flos": 2204652011520000.0,
"trial_name": null,
"trial_params": null
}