ben81828 commited on
Commit
51f8b15
·
verified ·
1 Parent(s): 2b7ec0b

Training in progress, step 750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:086f56d8253ad2de065083ad9ed3e401f206ca6169266148720dc9056e97e11d
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c65b6b24daabfe64ef0f6c6a0cc775ba02b62fb7a4c18bdac4a973c02aef24f
3
  size 29034840
last-checkpoint/global_step750/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:beb47aeb14a8461d8746421a15ccdc6228c1c958b75efa59b9309c850bc3df92
3
+ size 43429616
last-checkpoint/global_step750/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6636dd3c88527c8676a1c724e2fb7a0720bed30ea3ffbe86356ebc9fea193980
3
+ size 43429616
last-checkpoint/global_step750/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cfdedc3d830cac142e8e593bff3ac4c5b883c6c4aa3b08d2c0b666c1b9f94cc1
3
+ size 43429616
last-checkpoint/global_step750/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d48b4016ca4042b4b89b2e2a2f9d5265a3ac8b47e10f4d0fe183c7ef59fa4c94
3
+ size 43429616
last-checkpoint/global_step750/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b7ceefcf5f9aaf423f675a59b70c7f49dd63bf2969dd0f241c706afca330069
3
+ size 637299
last-checkpoint/global_step750/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88f0e9d3d0064652385de4668d7d1493464ea47cc90272985cc5f82378fe9407
3
+ size 637171
last-checkpoint/global_step750/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e2747d5b9a430ea06aa5e313062964f76ced72bafedfb86736e283a3440555b
3
+ size 637171
last-checkpoint/global_step750/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3df739d3279f7e7c62c77653c661bc9074781cffe48c2820c6dcb897ac09d46b
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step700
 
1
+ global_step750
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e7d74de51245105e1fbf57a6707ef3538b353952485508f6e2f8f74dc5d479d4
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:66b4ef73f3603a1b91082ee108fa8299ebe45fb3cdeec7d0bdca1982af5bf07d
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0617c9eb6cf7df57b2e0bb53cfe17c05f0910de56fe5b14427fe39ab54a44782
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:882eccb2a272cf97cd645050bd077c971e48e78584f717a1b1cc9b5f1c9326dc
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ed68a365057022897d9645ee60902a77102f43215dcdf2ddd5d3842b6a8446d8
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1626bac54f5806a604b7efdd749c5b65d63bbb40fc55c3744aae6130aa24f3de
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:63ebaa0c302cadbdfcd9f8ee2289e35ecf9c9fc8c9968fc0c05f100dac20c6b9
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40cf83ff997228172cf0b991f9d5209728ccf2f0a75841db5e31e647779a1ad2
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c2f570b3fae67de7e649855487e96bf548cf918d7ba3d40d1baf265735ed69a0
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:063f87ae0f4d3da8752b393e2d03482641fa84f93994bbbeba6ee3fa87c622b1
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7318872809410095,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-700",
4
- "epoch": 0.3605459696111254,
5
  "eval_steps": 50,
6
- "global_step": 700,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1253,11 +1253,100 @@
1253
  "eval_steps_per_second": 0.762,
1254
  "num_input_tokens_seen": 8187320,
1255
  "step": 700
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1256
  }
1257
  ],
1258
  "logging_steps": 5,
1259
  "max_steps": 3882,
1260
- "num_input_tokens_seen": 8187320,
1261
  "num_train_epochs": 2,
1262
  "save_steps": 50,
1263
  "stateful_callbacks": {
@@ -1272,7 +1361,7 @@
1272
  "attributes": {}
1273
  }
1274
  },
1275
- "total_flos": 540175003287552.0,
1276
  "train_batch_size": 1,
1277
  "trial_name": null,
1278
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7318872809410095,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-700",
4
+ "epoch": 0.38629925315477726,
5
  "eval_steps": 50,
6
+ "global_step": 750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1253
  "eval_steps_per_second": 0.762,
1254
  "num_input_tokens_seen": 8187320,
1255
  "step": 700
1256
+ },
1257
+ {
1258
+ "epoch": 0.3631212979654906,
1259
+ "grad_norm": 14.492396688166755,
1260
+ "learning_rate": 9.535282851249103e-05,
1261
+ "loss": 0.765,
1262
+ "num_input_tokens_seen": 8245776,
1263
+ "step": 705
1264
+ },
1265
+ {
1266
+ "epoch": 0.3656966263198558,
1267
+ "grad_norm": 16.054951836007135,
1268
+ "learning_rate": 9.526273465583646e-05,
1269
+ "loss": 0.7287,
1270
+ "num_input_tokens_seen": 8304280,
1271
+ "step": 710
1272
+ },
1273
+ {
1274
+ "epoch": 0.36827195467422097,
1275
+ "grad_norm": 21.4994447839661,
1276
+ "learning_rate": 9.517181924991716e-05,
1277
+ "loss": 0.758,
1278
+ "num_input_tokens_seen": 8362728,
1279
+ "step": 715
1280
+ },
1281
+ {
1282
+ "epoch": 0.37084728302858616,
1283
+ "grad_norm": 10.676012702912917,
1284
+ "learning_rate": 9.508008394490926e-05,
1285
+ "loss": 0.795,
1286
+ "num_input_tokens_seen": 8421224,
1287
+ "step": 720
1288
+ },
1289
+ {
1290
+ "epoch": 0.37342261138295135,
1291
+ "grad_norm": 9.802559177691224,
1292
+ "learning_rate": 9.498753040587066e-05,
1293
+ "loss": 0.6901,
1294
+ "num_input_tokens_seen": 8479720,
1295
+ "step": 725
1296
+ },
1297
+ {
1298
+ "epoch": 0.3759979397373165,
1299
+ "grad_norm": 12.345047855457121,
1300
+ "learning_rate": 9.48941603127107e-05,
1301
+ "loss": 0.7618,
1302
+ "num_input_tokens_seen": 8538192,
1303
+ "step": 730
1304
+ },
1305
+ {
1306
+ "epoch": 0.3785732680916817,
1307
+ "grad_norm": 9.115843124142248,
1308
+ "learning_rate": 9.479997536015977e-05,
1309
+ "loss": 0.7481,
1310
+ "num_input_tokens_seen": 8596664,
1311
+ "step": 735
1312
+ },
1313
+ {
1314
+ "epoch": 0.3811485964460469,
1315
+ "grad_norm": 13.754407712653018,
1316
+ "learning_rate": 9.47049772577385e-05,
1317
+ "loss": 0.746,
1318
+ "num_input_tokens_seen": 8655128,
1319
+ "step": 740
1320
+ },
1321
+ {
1322
+ "epoch": 0.38372392480041206,
1323
+ "grad_norm": 11.612003816357428,
1324
+ "learning_rate": 9.460916772972672e-05,
1325
+ "loss": 0.812,
1326
+ "num_input_tokens_seen": 8713624,
1327
+ "step": 745
1328
+ },
1329
+ {
1330
+ "epoch": 0.38629925315477726,
1331
+ "grad_norm": 27.606819697848053,
1332
+ "learning_rate": 9.451254851513222e-05,
1333
+ "loss": 0.6751,
1334
+ "num_input_tokens_seen": 8772104,
1335
+ "step": 750
1336
+ },
1337
+ {
1338
+ "epoch": 0.38629925315477726,
1339
+ "eval_loss": 0.732211709022522,
1340
+ "eval_runtime": 19.8828,
1341
+ "eval_samples_per_second": 3.018,
1342
+ "eval_steps_per_second": 0.754,
1343
+ "num_input_tokens_seen": 8772104,
1344
+ "step": 750
1345
  }
1346
  ],
1347
  "logging_steps": 5,
1348
  "max_steps": 3882,
1349
+ "num_input_tokens_seen": 8772104,
1350
  "num_train_epochs": 2,
1351
  "save_steps": 50,
1352
  "stateful_callbacks": {
 
1361
  "attributes": {}
1362
  }
1363
  },
1364
+ "total_flos": 578762919116800.0,
1365
  "train_batch_size": 1,
1366
  "trial_name": null,
1367
  "trial_params": null