ben81828 commited on
Commit
65e1f2d
·
verified ·
1 Parent(s): 3f5bab6

Training in progress, step 450, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d1ad4a76028a7f83ee44275d3e3a7c646bb4cd4ceb75431117f2d8b0f219b774
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7156b16f127d3e2296e6ada4ebf9cbb8c48381f16d088d0a874a522c40e99a16
3
  size 29034840
last-checkpoint/global_step450/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a06e1adc0a80d44504c1dee1034eca4ca2963ada9c46ca490451212b5e42fbd2
3
+ size 43429616
last-checkpoint/global_step450/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7851ab2d0a65baa1034c92407e075dd34b436bdc15606a42192d3eeb5625274e
3
+ size 43429616
last-checkpoint/global_step450/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:092821d92b2be112f83a63e8e11290ed76faf8f75d97035c702643fbdee6fc73
3
+ size 43429616
last-checkpoint/global_step450/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2c4afd4a5961b09abdb5fb5f9e5b29c3f2a8f6df9ebd068bb2b373e6cf53991
3
+ size 43429616
last-checkpoint/global_step450/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2c43b6581640ec9140a8e671944ff69d96a0284c4069ddbe5a73547523d050e
3
+ size 637299
last-checkpoint/global_step450/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e7b078653231149a903bc00280c61fc902a6ebe6078af02817b5efd13e9ef2b
3
+ size 637171
last-checkpoint/global_step450/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3014bb4485c273ec85ee6adab304455dcd75b99ca668f9c10e04f1ebc443fec8
3
+ size 637171
last-checkpoint/global_step450/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f23e7a4eb8e027c3a1e9ed72ab3097610a2c4faa00790d4b003da7cdb06018e2
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step400
 
1
+ global_step450
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:543ef05f530d40ee20b8d626b07a69b86597aca643e48897571062f973efe84f
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7dbc6521b0b64cb12d818506108fcf257a4089ca8a9b1e453776ed3e032e7176
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a23f732e43838ce0398d2636885ac16badbb9bcbc04d1406069ba3027bc5ae0
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b13e3da1b0679cab1bab94f893e385a9a224d3335b5a6f62602f33c2be88d03
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e10cce960e7068b051c05e35ed6160656be9091c63f13796ac2ed7e9c84e5a72
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a24f0e0f117b5a8236e0d12594c0c358f41ef00068d4460002e95ad1cc3cb1c
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b6f6049e212b1df5cefc5d834afcd8cc052c73f1457449e9fe8a38d514f54078
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e46e4eab6c4a25d84ad36ddf1357401788adeeb6388c03cefa35a63b52ee7610
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7779b5dd7047c5846ea43ed002061c3588a7e29e0d9f0ca42655eab3f8410e99
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:581bd51b94cb26100e335b86a6da2b6d11272f4e7dfeb8188f963bb8c6dcbb27
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.8932263255119324,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-300",
4
- "epoch": 0.20602626834921453,
5
  "eval_steps": 50,
6
- "global_step": 400,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -719,11 +719,100 @@
719
  "eval_steps_per_second": 0.757,
720
  "num_input_tokens_seen": 4678384,
721
  "step": 400
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
722
  }
723
  ],
724
  "logging_steps": 5,
725
  "max_steps": 3882,
726
- "num_input_tokens_seen": 4678384,
727
  "num_train_epochs": 2,
728
  "save_steps": 50,
729
  "stateful_callbacks": {
@@ -738,7 +827,7 @@
738
  "attributes": {}
739
  }
740
  },
741
- "total_flos": 308638862540800.0,
742
  "train_batch_size": 1,
743
  "trial_name": null,
744
  "trial_params": null
 
1
  {
2
  "best_metric": 0.8932263255119324,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-300",
4
+ "epoch": 0.23177955189286634,
5
  "eval_steps": 50,
6
+ "global_step": 450,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
719
  "eval_steps_per_second": 0.757,
720
  "num_input_tokens_seen": 4678384,
721
  "step": 400
722
+ },
723
+ {
724
+ "epoch": 0.2086015967035797,
725
+ "grad_norm": 0.4854922483196253,
726
+ "learning_rate": 9.920168738030807e-05,
727
+ "loss": 0.8951,
728
+ "num_input_tokens_seen": 4736904,
729
+ "step": 405
730
+ },
731
+ {
732
+ "epoch": 0.2111769250579449,
733
+ "grad_norm": 0.4537032287993717,
734
+ "learning_rate": 9.916332757010799e-05,
735
+ "loss": 0.9131,
736
+ "num_input_tokens_seen": 4795376,
737
+ "step": 410
738
+ },
739
+ {
740
+ "epoch": 0.21375225341231008,
741
+ "grad_norm": 0.46655140937482126,
742
+ "learning_rate": 9.912407541224213e-05,
743
+ "loss": 0.8923,
744
+ "num_input_tokens_seen": 4853880,
745
+ "step": 415
746
+ },
747
+ {
748
+ "epoch": 0.21632758176667524,
749
+ "grad_norm": 0.3608301258843965,
750
+ "learning_rate": 9.908393161916374e-05,
751
+ "loss": 0.9026,
752
+ "num_input_tokens_seen": 4912360,
753
+ "step": 420
754
+ },
755
+ {
756
+ "epoch": 0.21890291012104043,
757
+ "grad_norm": 0.5055648531803498,
758
+ "learning_rate": 9.904289691950979e-05,
759
+ "loss": 0.905,
760
+ "num_input_tokens_seen": 4970872,
761
+ "step": 425
762
+ },
763
+ {
764
+ "epoch": 0.22147823847540563,
765
+ "grad_norm": 0.3171606869940592,
766
+ "learning_rate": 9.900097205808789e-05,
767
+ "loss": 0.8941,
768
+ "num_input_tokens_seen": 5029304,
769
+ "step": 430
770
+ },
771
+ {
772
+ "epoch": 0.2240535668297708,
773
+ "grad_norm": 0.5798722428230844,
774
+ "learning_rate": 9.895815779586262e-05,
775
+ "loss": 0.9031,
776
+ "num_input_tokens_seen": 5087800,
777
+ "step": 435
778
+ },
779
+ {
780
+ "epoch": 0.22662889518413598,
781
+ "grad_norm": 0.5751083474979835,
782
+ "learning_rate": 9.891445490994182e-05,
783
+ "loss": 0.8975,
784
+ "num_input_tokens_seen": 5146312,
785
+ "step": 440
786
+ },
787
+ {
788
+ "epoch": 0.22920422353850115,
789
+ "grad_norm": 1.2316619525123293,
790
+ "learning_rate": 9.886986419356246e-05,
791
+ "loss": 0.901,
792
+ "num_input_tokens_seen": 5204800,
793
+ "step": 445
794
+ },
795
+ {
796
+ "epoch": 0.23177955189286634,
797
+ "grad_norm": 0.5953254293558816,
798
+ "learning_rate": 9.88243864560762e-05,
799
+ "loss": 0.9062,
800
+ "num_input_tokens_seen": 5263304,
801
+ "step": 450
802
+ },
803
+ {
804
+ "epoch": 0.23177955189286634,
805
+ "eval_loss": 0.9007609486579895,
806
+ "eval_runtime": 20.8827,
807
+ "eval_samples_per_second": 2.873,
808
+ "eval_steps_per_second": 0.718,
809
+ "num_input_tokens_seen": 5263304,
810
+ "step": 450
811
  }
812
  ],
813
  "logging_steps": 5,
814
  "max_steps": 3882,
815
+ "num_input_tokens_seen": 5263304,
816
  "num_train_epochs": 2,
817
  "save_steps": 50,
818
  "stateful_callbacks": {
 
827
  "attributes": {}
828
  }
829
  },
830
+ "total_flos": 347233811955712.0,
831
  "train_batch_size": 1,
832
  "trial_name": null,
833
  "trial_params": null