|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 6.0, |
|
"eval_steps": 500, |
|
"global_step": 62268, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.05, |
|
"learning_rate": 4.95985096678872e-05, |
|
"loss": 19.0199, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"learning_rate": 4.91970193357744e-05, |
|
"loss": 2.0644, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"learning_rate": 4.879552900366159e-05, |
|
"loss": 0.8825, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"learning_rate": 4.839403867154879e-05, |
|
"loss": 0.6416, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"learning_rate": 4.799254833943599e-05, |
|
"loss": 0.5382, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"learning_rate": 4.759105800732319e-05, |
|
"loss": 0.4854, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"learning_rate": 4.7189567675210385e-05, |
|
"loss": 0.4543, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"learning_rate": 4.678807734309758e-05, |
|
"loss": 0.4062, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"learning_rate": 4.6386587010984776e-05, |
|
"loss": 0.3786, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"learning_rate": 4.5985096678871974e-05, |
|
"loss": 0.3586, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.53, |
|
"learning_rate": 4.558360634675917e-05, |
|
"loss": 0.3445, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.58, |
|
"learning_rate": 4.518211601464637e-05, |
|
"loss": 0.3206, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"learning_rate": 4.478062568253357e-05, |
|
"loss": 0.301, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.67, |
|
"learning_rate": 4.437913535042076e-05, |
|
"loss": 0.2895, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"learning_rate": 4.397764501830796e-05, |
|
"loss": 0.2801, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.77, |
|
"learning_rate": 4.357615468619516e-05, |
|
"loss": 0.2765, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.82, |
|
"learning_rate": 4.317466435408236e-05, |
|
"loss": 0.2644, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"learning_rate": 4.2773174021969556e-05, |
|
"loss": 0.2631, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"learning_rate": 4.237168368985675e-05, |
|
"loss": 0.2477, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"learning_rate": 4.1970193357743946e-05, |
|
"loss": 0.2512, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"eval_loss": 0.19760668277740479, |
|
"eval_runtime": 169.1284, |
|
"eval_samples_per_second": 66.039, |
|
"eval_steps_per_second": 8.26, |
|
"step": 10378 |
|
}, |
|
{ |
|
"epoch": 1.01, |
|
"learning_rate": 4.1568703025631145e-05, |
|
"loss": 0.2432, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 1.06, |
|
"learning_rate": 4.116721269351834e-05, |
|
"loss": 0.2329, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"learning_rate": 4.076572236140554e-05, |
|
"loss": 0.2353, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"learning_rate": 4.0364232029292733e-05, |
|
"loss": 0.2274, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"learning_rate": 3.996274169717993e-05, |
|
"loss": 0.2215, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 1.25, |
|
"learning_rate": 3.956125136506713e-05, |
|
"loss": 0.2187, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 1.3, |
|
"learning_rate": 3.915976103295433e-05, |
|
"loss": 0.2151, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"learning_rate": 3.875827070084153e-05, |
|
"loss": 0.218, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"learning_rate": 3.8356780368728726e-05, |
|
"loss": 0.2106, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 1.45, |
|
"learning_rate": 3.795529003661592e-05, |
|
"loss": 0.2109, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 1.49, |
|
"learning_rate": 3.7553799704503117e-05, |
|
"loss": 0.2061, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 1.54, |
|
"learning_rate": 3.7152309372390315e-05, |
|
"loss": 0.204, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"learning_rate": 3.6750819040277514e-05, |
|
"loss": 0.2047, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"learning_rate": 3.634932870816471e-05, |
|
"loss": 0.2028, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 1.69, |
|
"learning_rate": 3.5947838376051904e-05, |
|
"loss": 0.1929, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 1.73, |
|
"learning_rate": 3.55463480439391e-05, |
|
"loss": 0.1984, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 1.78, |
|
"learning_rate": 3.51448577118263e-05, |
|
"loss": 0.1901, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"learning_rate": 3.47433673797135e-05, |
|
"loss": 0.1974, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"learning_rate": 3.43418770476007e-05, |
|
"loss": 0.2006, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 1.93, |
|
"learning_rate": 3.394038671548789e-05, |
|
"loss": 0.1885, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"learning_rate": 3.353889638337509e-05, |
|
"loss": 0.1827, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"eval_loss": 0.1765761375427246, |
|
"eval_runtime": 169.0515, |
|
"eval_samples_per_second": 66.069, |
|
"eval_steps_per_second": 8.264, |
|
"step": 20756 |
|
}, |
|
{ |
|
"epoch": 2.02, |
|
"learning_rate": 3.313740605126229e-05, |
|
"loss": 0.1909, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"learning_rate": 3.2735915719149485e-05, |
|
"loss": 0.1796, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"learning_rate": 3.2334425387036684e-05, |
|
"loss": 0.1785, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 2.17, |
|
"learning_rate": 3.193293505492388e-05, |
|
"loss": 0.1767, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"learning_rate": 3.1531444722811074e-05, |
|
"loss": 0.1764, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 2.26, |
|
"learning_rate": 3.112995439069827e-05, |
|
"loss": 0.1777, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"learning_rate": 3.072846405858547e-05, |
|
"loss": 0.1722, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"learning_rate": 3.032697372647267e-05, |
|
"loss": 0.1759, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 2.41, |
|
"learning_rate": 2.9925483394359865e-05, |
|
"loss": 0.172, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"learning_rate": 2.9523993062247064e-05, |
|
"loss": 0.172, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"learning_rate": 2.912250273013426e-05, |
|
"loss": 0.1763, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"learning_rate": 2.8721012398021457e-05, |
|
"loss": 0.1712, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"learning_rate": 2.8319522065908656e-05, |
|
"loss": 0.1666, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 2.65, |
|
"learning_rate": 2.791803173379585e-05, |
|
"loss": 0.1704, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"learning_rate": 2.751654140168305e-05, |
|
"loss": 0.1695, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"learning_rate": 2.7115051069570248e-05, |
|
"loss": 0.1643, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"learning_rate": 2.6713560737457443e-05, |
|
"loss": 0.1666, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"learning_rate": 2.6312070405344642e-05, |
|
"loss": 0.1711, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 2.89, |
|
"learning_rate": 2.5910580073231837e-05, |
|
"loss": 0.1662, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"learning_rate": 2.5509089741119035e-05, |
|
"loss": 0.1663, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"learning_rate": 2.5107599409006234e-05, |
|
"loss": 0.1661, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"eval_loss": 0.1699543297290802, |
|
"eval_runtime": 168.4223, |
|
"eval_samples_per_second": 66.315, |
|
"eval_steps_per_second": 8.295, |
|
"step": 31134 |
|
}, |
|
{ |
|
"epoch": 3.04, |
|
"learning_rate": 2.470610907689343e-05, |
|
"loss": 0.1641, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 3.08, |
|
"learning_rate": 2.4304618744780628e-05, |
|
"loss": 0.1577, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 3.13, |
|
"learning_rate": 2.3903128412667826e-05, |
|
"loss": 0.1588, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 3.18, |
|
"learning_rate": 2.350163808055502e-05, |
|
"loss": 0.1569, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 3.23, |
|
"learning_rate": 2.310014774844222e-05, |
|
"loss": 0.1616, |
|
"step": 33500 |
|
}, |
|
{ |
|
"epoch": 3.28, |
|
"learning_rate": 2.2698657416329415e-05, |
|
"loss": 0.1531, |
|
"step": 34000 |
|
}, |
|
{ |
|
"epoch": 3.32, |
|
"learning_rate": 2.2297167084216614e-05, |
|
"loss": 0.1642, |
|
"step": 34500 |
|
}, |
|
{ |
|
"epoch": 3.37, |
|
"learning_rate": 2.1895676752103812e-05, |
|
"loss": 0.1571, |
|
"step": 35000 |
|
}, |
|
{ |
|
"epoch": 3.42, |
|
"learning_rate": 2.1494186419991007e-05, |
|
"loss": 0.1525, |
|
"step": 35500 |
|
}, |
|
{ |
|
"epoch": 3.47, |
|
"learning_rate": 2.1092696087878206e-05, |
|
"loss": 0.1509, |
|
"step": 36000 |
|
}, |
|
{ |
|
"epoch": 3.52, |
|
"learning_rate": 2.06912057557654e-05, |
|
"loss": 0.1537, |
|
"step": 36500 |
|
}, |
|
{ |
|
"epoch": 3.57, |
|
"learning_rate": 2.02897154236526e-05, |
|
"loss": 0.1554, |
|
"step": 37000 |
|
}, |
|
{ |
|
"epoch": 3.61, |
|
"learning_rate": 1.9888225091539798e-05, |
|
"loss": 0.1512, |
|
"step": 37500 |
|
}, |
|
{ |
|
"epoch": 3.66, |
|
"learning_rate": 1.9486734759426993e-05, |
|
"loss": 0.151, |
|
"step": 38000 |
|
}, |
|
{ |
|
"epoch": 3.71, |
|
"learning_rate": 1.9085244427314192e-05, |
|
"loss": 0.1519, |
|
"step": 38500 |
|
}, |
|
{ |
|
"epoch": 3.76, |
|
"learning_rate": 1.868375409520139e-05, |
|
"loss": 0.1534, |
|
"step": 39000 |
|
}, |
|
{ |
|
"epoch": 3.81, |
|
"learning_rate": 1.8282263763088585e-05, |
|
"loss": 0.1551, |
|
"step": 39500 |
|
}, |
|
{ |
|
"epoch": 3.85, |
|
"learning_rate": 1.7880773430975784e-05, |
|
"loss": 0.1506, |
|
"step": 40000 |
|
}, |
|
{ |
|
"epoch": 3.9, |
|
"learning_rate": 1.747928309886298e-05, |
|
"loss": 0.1479, |
|
"step": 40500 |
|
}, |
|
{ |
|
"epoch": 3.95, |
|
"learning_rate": 1.7077792766750178e-05, |
|
"loss": 0.1485, |
|
"step": 41000 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"learning_rate": 1.6676302434637376e-05, |
|
"loss": 0.1545, |
|
"step": 41500 |
|
}, |
|
{ |
|
"epoch": 4.0, |
|
"eval_loss": 0.16708502173423767, |
|
"eval_runtime": 168.3042, |
|
"eval_samples_per_second": 66.362, |
|
"eval_steps_per_second": 8.3, |
|
"step": 41512 |
|
}, |
|
{ |
|
"epoch": 4.05, |
|
"learning_rate": 1.627481210252457e-05, |
|
"loss": 0.1467, |
|
"step": 42000 |
|
}, |
|
{ |
|
"epoch": 4.1, |
|
"learning_rate": 1.587332177041177e-05, |
|
"loss": 0.1478, |
|
"step": 42500 |
|
}, |
|
{ |
|
"epoch": 4.14, |
|
"learning_rate": 1.547183143829897e-05, |
|
"loss": 0.1452, |
|
"step": 43000 |
|
}, |
|
{ |
|
"epoch": 4.19, |
|
"learning_rate": 1.5070341106186164e-05, |
|
"loss": 0.1464, |
|
"step": 43500 |
|
}, |
|
{ |
|
"epoch": 4.24, |
|
"learning_rate": 1.4668850774073362e-05, |
|
"loss": 0.15, |
|
"step": 44000 |
|
}, |
|
{ |
|
"epoch": 4.29, |
|
"learning_rate": 1.4267360441960559e-05, |
|
"loss": 0.1433, |
|
"step": 44500 |
|
}, |
|
{ |
|
"epoch": 4.34, |
|
"learning_rate": 1.3865870109847756e-05, |
|
"loss": 0.1431, |
|
"step": 45000 |
|
}, |
|
{ |
|
"epoch": 4.38, |
|
"learning_rate": 1.3464379777734953e-05, |
|
"loss": 0.1424, |
|
"step": 45500 |
|
}, |
|
{ |
|
"epoch": 4.43, |
|
"learning_rate": 1.3062889445622151e-05, |
|
"loss": 0.14, |
|
"step": 46000 |
|
}, |
|
{ |
|
"epoch": 4.48, |
|
"learning_rate": 1.2661399113509348e-05, |
|
"loss": 0.1443, |
|
"step": 46500 |
|
}, |
|
{ |
|
"epoch": 4.53, |
|
"learning_rate": 1.2259908781396545e-05, |
|
"loss": 0.1461, |
|
"step": 47000 |
|
}, |
|
{ |
|
"epoch": 4.58, |
|
"learning_rate": 1.1858418449283742e-05, |
|
"loss": 0.1431, |
|
"step": 47500 |
|
}, |
|
{ |
|
"epoch": 4.63, |
|
"learning_rate": 1.145692811717094e-05, |
|
"loss": 0.1403, |
|
"step": 48000 |
|
}, |
|
{ |
|
"epoch": 4.67, |
|
"learning_rate": 1.1055437785058137e-05, |
|
"loss": 0.1442, |
|
"step": 48500 |
|
}, |
|
{ |
|
"epoch": 4.72, |
|
"learning_rate": 1.0653947452945334e-05, |
|
"loss": 0.1468, |
|
"step": 49000 |
|
}, |
|
{ |
|
"epoch": 4.77, |
|
"learning_rate": 1.0252457120832531e-05, |
|
"loss": 0.1483, |
|
"step": 49500 |
|
}, |
|
{ |
|
"epoch": 4.82, |
|
"learning_rate": 9.85096678871973e-06, |
|
"loss": 0.1402, |
|
"step": 50000 |
|
}, |
|
{ |
|
"epoch": 4.87, |
|
"learning_rate": 9.449476456606926e-06, |
|
"loss": 0.1428, |
|
"step": 50500 |
|
}, |
|
{ |
|
"epoch": 4.91, |
|
"learning_rate": 9.047986124494123e-06, |
|
"loss": 0.1447, |
|
"step": 51000 |
|
}, |
|
{ |
|
"epoch": 4.96, |
|
"learning_rate": 8.64649579238132e-06, |
|
"loss": 0.1472, |
|
"step": 51500 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"eval_loss": 0.16224625706672668, |
|
"eval_runtime": 168.3955, |
|
"eval_samples_per_second": 66.326, |
|
"eval_steps_per_second": 8.296, |
|
"step": 51890 |
|
}, |
|
{ |
|
"epoch": 5.01, |
|
"learning_rate": 8.245005460268517e-06, |
|
"loss": 0.1418, |
|
"step": 52000 |
|
}, |
|
{ |
|
"epoch": 5.06, |
|
"learning_rate": 7.843515128155714e-06, |
|
"loss": 0.1358, |
|
"step": 52500 |
|
}, |
|
{ |
|
"epoch": 5.11, |
|
"learning_rate": 7.442024796042911e-06, |
|
"loss": 0.1332, |
|
"step": 53000 |
|
}, |
|
{ |
|
"epoch": 5.16, |
|
"learning_rate": 7.040534463930108e-06, |
|
"loss": 0.1417, |
|
"step": 53500 |
|
}, |
|
{ |
|
"epoch": 5.2, |
|
"learning_rate": 6.639044131817306e-06, |
|
"loss": 0.1368, |
|
"step": 54000 |
|
}, |
|
{ |
|
"epoch": 5.25, |
|
"learning_rate": 6.2375537997045036e-06, |
|
"loss": 0.1422, |
|
"step": 54500 |
|
}, |
|
{ |
|
"epoch": 5.3, |
|
"learning_rate": 5.8360634675917e-06, |
|
"loss": 0.1329, |
|
"step": 55000 |
|
}, |
|
{ |
|
"epoch": 5.35, |
|
"learning_rate": 5.434573135478898e-06, |
|
"loss": 0.1347, |
|
"step": 55500 |
|
}, |
|
{ |
|
"epoch": 5.4, |
|
"learning_rate": 5.033082803366095e-06, |
|
"loss": 0.1468, |
|
"step": 56000 |
|
}, |
|
{ |
|
"epoch": 5.44, |
|
"learning_rate": 4.631592471253293e-06, |
|
"loss": 0.1348, |
|
"step": 56500 |
|
}, |
|
{ |
|
"epoch": 5.49, |
|
"learning_rate": 4.2301021391404895e-06, |
|
"loss": 0.14, |
|
"step": 57000 |
|
}, |
|
{ |
|
"epoch": 5.54, |
|
"learning_rate": 3.828611807027687e-06, |
|
"loss": 0.1414, |
|
"step": 57500 |
|
}, |
|
{ |
|
"epoch": 5.59, |
|
"learning_rate": 3.4271214749148845e-06, |
|
"loss": 0.1415, |
|
"step": 58000 |
|
}, |
|
{ |
|
"epoch": 5.64, |
|
"learning_rate": 3.0256311428020817e-06, |
|
"loss": 0.1346, |
|
"step": 58500 |
|
}, |
|
{ |
|
"epoch": 5.69, |
|
"learning_rate": 2.624140810689279e-06, |
|
"loss": 0.1375, |
|
"step": 59000 |
|
}, |
|
{ |
|
"epoch": 5.73, |
|
"learning_rate": 2.2226504785764763e-06, |
|
"loss": 0.1379, |
|
"step": 59500 |
|
}, |
|
{ |
|
"epoch": 5.78, |
|
"learning_rate": 1.8211601464636733e-06, |
|
"loss": 0.1365, |
|
"step": 60000 |
|
}, |
|
{ |
|
"epoch": 5.83, |
|
"learning_rate": 1.4196698143508704e-06, |
|
"loss": 0.1462, |
|
"step": 60500 |
|
}, |
|
{ |
|
"epoch": 5.88, |
|
"learning_rate": 1.0181794822380677e-06, |
|
"loss": 0.1356, |
|
"step": 61000 |
|
}, |
|
{ |
|
"epoch": 5.93, |
|
"learning_rate": 6.166891501252649e-07, |
|
"loss": 0.137, |
|
"step": 61500 |
|
}, |
|
{ |
|
"epoch": 5.97, |
|
"learning_rate": 2.1519881801246228e-07, |
|
"loss": 0.1429, |
|
"step": 62000 |
|
}, |
|
{ |
|
"epoch": 6.0, |
|
"eval_loss": 0.1622563898563385, |
|
"eval_runtime": 168.6575, |
|
"eval_samples_per_second": 66.223, |
|
"eval_steps_per_second": 8.283, |
|
"step": 62268 |
|
} |
|
], |
|
"logging_steps": 500, |
|
"max_steps": 62268, |
|
"num_train_epochs": 6, |
|
"save_steps": 500, |
|
"total_flos": 2.6337445102485504e+17, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|