Training in progress, step 2600, checkpoint
Browse files- last-checkpoint/adapter_model.safetensors +1 -1
- last-checkpoint/global_step2599/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2599/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2599/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2599/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step2599/zero_pp_rank_0_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step2599/zero_pp_rank_1_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step2599/zero_pp_rank_2_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step2599/zero_pp_rank_3_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/latest +1 -1
- last-checkpoint/rng_state_0.pth +1 -1
- last-checkpoint/rng_state_1.pth +1 -1
- last-checkpoint/rng_state_2.pth +1 -1
- last-checkpoint/rng_state_3.pth +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +95 -6
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 29034840
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:822e6dcdbe992039c470c7d5345a56c4a6d718bf3296adb52b926cfb46d4ab8e
|
3 |
size 29034840
|
last-checkpoint/global_step2599/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e095771b1dd0a77832675a32c1e0adba637ce893192708948322e4ae0e108d85
|
3 |
+
size 43429616
|
last-checkpoint/global_step2599/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d8bfec6d94c896829153a3df913f4031650e5f5e6cd918e31315a75f8de3c9c
|
3 |
+
size 43429616
|
last-checkpoint/global_step2599/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:011fb90eb0ac52df7b6b124720eba263bde86534222e1c70c19d554967ee6a75
|
3 |
+
size 43429616
|
last-checkpoint/global_step2599/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8af9b47439108751362a728f0e1f6ff8284c1aac5c0f85286e109f3864ad8508
|
3 |
+
size 43429616
|
last-checkpoint/global_step2599/zero_pp_rank_0_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a33d0c44522250f6271beb76c30d602be4c8a12e9374721e9ac3f7ed9aa4cdf0
|
3 |
+
size 637299
|
last-checkpoint/global_step2599/zero_pp_rank_1_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1f057ab5447205b9eb14de207e7b2fd1b440d8f383f812bb4efbe5197c5bf59
|
3 |
+
size 637171
|
last-checkpoint/global_step2599/zero_pp_rank_2_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4bffd784ee7b57957293b50b7f079fdfd39c1d272fb838f2f34e66e8e8a1176d
|
3 |
+
size 637171
|
last-checkpoint/global_step2599/zero_pp_rank_3_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2fc176fcd1b7abc3279e54d33aef1bf0c3f4582cf78c44322368da5ceeeb2a3
|
3 |
+
size 637171
|
last-checkpoint/latest
CHANGED
@@ -1 +1 @@
|
|
1 |
-
|
|
|
1 |
+
global_step2599
|
last-checkpoint/rng_state_0.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7bf1e520c0ffd0912cdaf36292baf0834dd187b6292436dd343acbd7d39f37df
|
3 |
size 15024
|
last-checkpoint/rng_state_1.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:480a3937172137fc1b4a3886bc929d7dd5dd0507aebd6b20448e5d03ffe9f33f
|
3 |
size 15024
|
last-checkpoint/rng_state_2.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9cc7d6afeed03ecf805019b6cd001604bbcd5e4ef0cbc38e4dc6f5587e48dbf2
|
3 |
size 15024
|
last-checkpoint/rng_state_3.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 15024
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:48c0ace59112a18d7f177314d1a9eb2b6b894f2028df1dd368b8c67c8732e18d
|
3 |
size 15024
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2794fc624dc0bb5af51b88db4a8a47eb798dc318755b3b6d0e646a1f265656ea
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"best_metric": 0.
|
3 |
-
"best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-
|
4 |
-
"epoch": 1.
|
5 |
"eval_steps": 50,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -4546,11 +4546,100 @@
|
|
4546 |
"eval_steps_per_second": 0.77,
|
4547 |
"num_input_tokens_seen": 29818856,
|
4548 |
"step": 2550
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
4549 |
}
|
4550 |
],
|
4551 |
"logging_steps": 5,
|
4552 |
"max_steps": 3882,
|
4553 |
-
"num_input_tokens_seen":
|
4554 |
"num_train_epochs": 2,
|
4555 |
"save_steps": 50,
|
4556 |
"stateful_callbacks": {
|
@@ -4565,7 +4654,7 @@
|
|
4565 |
"attributes": {}
|
4566 |
}
|
4567 |
},
|
4568 |
-
"total_flos":
|
4569 |
"train_batch_size": 1,
|
4570 |
"trial_name": null,
|
4571 |
"trial_params": null
|
|
|
1 |
{
|
2 |
+
"best_metric": 0.20262545347213745,
|
3 |
+
"best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-2600",
|
4 |
+
"epoch": 1.338913211434458,
|
5 |
"eval_steps": 50,
|
6 |
+
"global_step": 2600,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
4546 |
"eval_steps_per_second": 0.77,
|
4547 |
"num_input_tokens_seen": 29818856,
|
4548 |
"step": 2550
|
4549 |
+
},
|
4550 |
+
{
|
4551 |
+
"epoch": 1.3157352562451712,
|
4552 |
+
"grad_norm": 5.864909751734038,
|
4553 |
+
"learning_rate": 2.8698700556002178e-05,
|
4554 |
+
"loss": 0.2231,
|
4555 |
+
"num_input_tokens_seen": 29877288,
|
4556 |
+
"step": 2555
|
4557 |
+
},
|
4558 |
+
{
|
4559 |
+
"epoch": 1.3183105845995364,
|
4560 |
+
"grad_norm": 11.12053960539336,
|
4561 |
+
"learning_rate": 2.8506174548135695e-05,
|
4562 |
+
"loss": 0.3007,
|
4563 |
+
"num_input_tokens_seen": 29935776,
|
4564 |
+
"step": 2560
|
4565 |
+
},
|
4566 |
+
{
|
4567 |
+
"epoch": 1.3208859129539017,
|
4568 |
+
"grad_norm": 5.495356583672569,
|
4569 |
+
"learning_rate": 2.8314038667746732e-05,
|
4570 |
+
"loss": 0.1895,
|
4571 |
+
"num_input_tokens_seen": 29994232,
|
4572 |
+
"step": 2565
|
4573 |
+
},
|
4574 |
+
{
|
4575 |
+
"epoch": 1.3234612413082667,
|
4576 |
+
"grad_norm": 11.482142819244615,
|
4577 |
+
"learning_rate": 2.81222964022315e-05,
|
4578 |
+
"loss": 0.2728,
|
4579 |
+
"num_input_tokens_seen": 30052760,
|
4580 |
+
"step": 2570
|
4581 |
+
},
|
4582 |
+
{
|
4583 |
+
"epoch": 1.326036569662632,
|
4584 |
+
"grad_norm": 8.217608228490523,
|
4585 |
+
"learning_rate": 2.793095123184193e-05,
|
4586 |
+
"loss": 0.2827,
|
4587 |
+
"num_input_tokens_seen": 30111200,
|
4588 |
+
"step": 2575
|
4589 |
+
},
|
4590 |
+
{
|
4591 |
+
"epoch": 1.3286118980169972,
|
4592 |
+
"grad_norm": 5.535041581031614,
|
4593 |
+
"learning_rate": 2.77400066296223e-05,
|
4594 |
+
"loss": 0.2191,
|
4595 |
+
"num_input_tokens_seen": 30169680,
|
4596 |
+
"step": 2580
|
4597 |
+
},
|
4598 |
+
{
|
4599 |
+
"epoch": 1.3311872263713624,
|
4600 |
+
"grad_norm": 6.639368336077178,
|
4601 |
+
"learning_rate": 2.7549466061346386e-05,
|
4602 |
+
"loss": 0.2317,
|
4603 |
+
"num_input_tokens_seen": 30228176,
|
4604 |
+
"step": 2585
|
4605 |
+
},
|
4606 |
+
{
|
4607 |
+
"epoch": 1.3337625547257275,
|
4608 |
+
"grad_norm": 5.625387657466934,
|
4609 |
+
"learning_rate": 2.7359332985454443e-05,
|
4610 |
+
"loss": 0.2293,
|
4611 |
+
"num_input_tokens_seen": 30286656,
|
4612 |
+
"step": 2590
|
4613 |
+
},
|
4614 |
+
{
|
4615 |
+
"epoch": 1.3363378830800927,
|
4616 |
+
"grad_norm": 3.631245501478199,
|
4617 |
+
"learning_rate": 2.716961085299049e-05,
|
4618 |
+
"loss": 0.1903,
|
4619 |
+
"num_input_tokens_seen": 30345160,
|
4620 |
+
"step": 2595
|
4621 |
+
},
|
4622 |
+
{
|
4623 |
+
"epoch": 1.338913211434458,
|
4624 |
+
"grad_norm": 11.391263066967507,
|
4625 |
+
"learning_rate": 2.698030310753958e-05,
|
4626 |
+
"loss": 0.224,
|
4627 |
+
"num_input_tokens_seen": 30403640,
|
4628 |
+
"step": 2600
|
4629 |
+
},
|
4630 |
+
{
|
4631 |
+
"epoch": 1.338913211434458,
|
4632 |
+
"eval_loss": 0.20262545347213745,
|
4633 |
+
"eval_runtime": 19.6125,
|
4634 |
+
"eval_samples_per_second": 3.059,
|
4635 |
+
"eval_steps_per_second": 0.765,
|
4636 |
+
"num_input_tokens_seen": 30403640,
|
4637 |
+
"step": 2600
|
4638 |
}
|
4639 |
],
|
4640 |
"logging_steps": 5,
|
4641 |
"max_steps": 3882,
|
4642 |
+
"num_input_tokens_seen": 30403640,
|
4643 |
"num_train_epochs": 2,
|
4644 |
"save_steps": 50,
|
4645 |
"stateful_callbacks": {
|
|
|
4654 |
"attributes": {}
|
4655 |
}
|
4656 |
},
|
4657 |
+
"total_flos": 2006172016574464.0,
|
4658 |
"train_batch_size": 1,
|
4659 |
"trial_name": null,
|
4660 |
"trial_params": null
|