ben81828 commited on
Commit
c3ef6fa
·
verified ·
1 Parent(s): 4ff6cb2

Training in progress, step 2750, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:493ef098b4e142b34c8f8ffa03b228d10404d9550092e5415ecc2ddce2804e56
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dcf51604c7fbe68fb5bec81cac3b4632fb619bd371e32820e3fd872616836861
3
  size 29034840
last-checkpoint/global_step2749/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16f749fa7242f50a15167a3cfc8f3f0ca21db7ef95db11825b34fe8477b9894f
3
+ size 43429616
last-checkpoint/global_step2749/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e282acf71b51832288b6023e31e80ec7d4a33addbf3307769341a67a260a5a8b
3
+ size 43429616
last-checkpoint/global_step2749/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05689186bf64e352efea034b0526442cf4788fb5ca574778846dbd5212a52ac4
3
+ size 43429616
last-checkpoint/global_step2749/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88284573bab80830396943150f719948ed82a68f579a0cb3e5bb27493aaf54b8
3
+ size 43429616
last-checkpoint/global_step2749/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1051febab85db07da6f2d5583551a973689717beb73a989075986b4a249b7e3d
3
+ size 637299
last-checkpoint/global_step2749/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07b952e7c1470e1848899b6793f447e4d319957d8b2072e541dcb23e86436686
3
+ size 637171
last-checkpoint/global_step2749/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:162e2a9507b0126ea6d58b1dff2e6435a0794913e45421499748680c5934a764
3
+ size 637171
last-checkpoint/global_step2749/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db1b6b12ab5689d184e76c5f3f7284d49c453eb7a4a7f2e7bb88f4bc020c628a
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step2699
 
1
+ global_step2749
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d8d3c7739f9787ea797b86ff1b3a51f9e68197835ba3178915a8a77558f67fc
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49db5a9fd0c84d580c671e52905ebeffc155b36537e76ff966d2e82906708999
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a22a57799bc43e59db67d9a787ed73040020c5f35990602033f4dab1318787d7
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8ca224562d8d97aaa131b3516288bb99f68d7dcf62170494326662bda0bb206
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29a624b936b77a04d6bfb6940acdd65a710bf39452e419e7ddb5c40fb2261072
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e86a11c89dba78d60e1b2a1855a651b90a5a22ef131ce65d26af83668c154e
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a79306817d4440cd621149537e8cf216b60f847fc6f9531a6147426aa02bb07
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3dd63019c923e9692431619aced46b91aaf3fd22e1c22ec0a64347f2fe635a0e
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55fc50c620128befbe42c0ee9c82ae426d9d0e14e63b4929582516d207937ad9
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2d5206e8eee1fd5c926f6f8750c100fc15237f73c48e6ae482a3b1e4fadc9091
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.18780523538589478,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-2650",
4
- "epoch": 1.3904197785217616,
5
  "eval_steps": 50,
6
- "global_step": 2700,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -4813,11 +4813,100 @@
4813
  "eval_steps_per_second": 0.774,
4814
  "num_input_tokens_seen": 31573240,
4815
  "step": 2700
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4816
  }
4817
  ],
4818
  "logging_steps": 5,
4819
  "max_steps": 3882,
4820
- "num_input_tokens_seen": 31573240,
4821
  "num_train_epochs": 2,
4822
  "save_steps": 50,
4823
  "stateful_callbacks": {
@@ -4832,7 +4921,7 @@
4832
  "attributes": {}
4833
  }
4834
  },
4835
- "total_flos": 2083346708430848.0,
4836
  "train_batch_size": 1,
4837
  "trial_name": null,
4838
  "trial_params": null
 
1
  {
2
  "best_metric": 0.18780523538589478,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-2650",
4
+ "epoch": 1.4161730620654134,
5
  "eval_steps": 50,
6
+ "global_step": 2750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
4813
  "eval_steps_per_second": 0.774,
4814
  "num_input_tokens_seen": 31573240,
4815
  "step": 2700
4816
+ },
4817
+ {
4818
+ "epoch": 1.3929951068761266,
4819
+ "grad_norm": 23.383619982845065,
4820
+ "learning_rate": 2.3106579887919894e-05,
4821
+ "loss": 0.2339,
4822
+ "num_input_tokens_seen": 31631688,
4823
+ "step": 2705
4824
+ },
4825
+ {
4826
+ "epoch": 1.3955704352304918,
4827
+ "grad_norm": 4.588108718468969,
4828
+ "learning_rate": 2.2927244003753694e-05,
4829
+ "loss": 0.2197,
4830
+ "num_input_tokens_seen": 31690200,
4831
+ "step": 2710
4832
+ },
4833
+ {
4834
+ "epoch": 1.398145763584857,
4835
+ "grad_norm": 3.4659557649616284,
4836
+ "learning_rate": 2.2748399508429834e-05,
4837
+ "loss": 0.1999,
4838
+ "num_input_tokens_seen": 31748664,
4839
+ "step": 2715
4840
+ },
4841
+ {
4842
+ "epoch": 1.4007210919392223,
4843
+ "grad_norm": 5.297859489588456,
4844
+ "learning_rate": 2.2570049648096902e-05,
4845
+ "loss": 0.2757,
4846
+ "num_input_tokens_seen": 31807176,
4847
+ "step": 2720
4848
+ },
4849
+ {
4850
+ "epoch": 1.4032964202935874,
4851
+ "grad_norm": 8.718394323594142,
4852
+ "learning_rate": 2.239219765992557e-05,
4853
+ "loss": 0.2128,
4854
+ "num_input_tokens_seen": 31865696,
4855
+ "step": 2725
4856
+ },
4857
+ {
4858
+ "epoch": 1.4058717486479526,
4859
+ "grad_norm": 11.91241028531398,
4860
+ "learning_rate": 2.221484677204978e-05,
4861
+ "loss": 0.2657,
4862
+ "num_input_tokens_seen": 31924200,
4863
+ "step": 2730
4864
+ },
4865
+ {
4866
+ "epoch": 1.4084470770023179,
4867
+ "grad_norm": 6.407371884386196,
4868
+ "learning_rate": 2.203800020350819e-05,
4869
+ "loss": 0.2349,
4870
+ "num_input_tokens_seen": 31982712,
4871
+ "step": 2735
4872
+ },
4873
+ {
4874
+ "epoch": 1.4110224053566829,
4875
+ "grad_norm": 4.998470631619502,
4876
+ "learning_rate": 2.1861661164185652e-05,
4877
+ "loss": 0.2188,
4878
+ "num_input_tokens_seen": 32041176,
4879
+ "step": 2740
4880
+ },
4881
+ {
4882
+ "epoch": 1.4135977337110481,
4883
+ "grad_norm": 6.503122447364567,
4884
+ "learning_rate": 2.1685832854755115e-05,
4885
+ "loss": 0.2531,
4886
+ "num_input_tokens_seen": 32099656,
4887
+ "step": 2745
4888
+ },
4889
+ {
4890
+ "epoch": 1.4161730620654134,
4891
+ "grad_norm": 3.8963890551505123,
4892
+ "learning_rate": 2.1510518466619334e-05,
4893
+ "loss": 0.1416,
4894
+ "num_input_tokens_seen": 32158144,
4895
+ "step": 2750
4896
+ },
4897
+ {
4898
+ "epoch": 1.4161730620654134,
4899
+ "eval_loss": 0.24715903401374817,
4900
+ "eval_runtime": 19.5042,
4901
+ "eval_samples_per_second": 3.076,
4902
+ "eval_steps_per_second": 0.769,
4903
+ "num_input_tokens_seen": 32158144,
4904
+ "step": 2750
4905
  }
4906
  ],
4907
  "logging_steps": 5,
4908
  "max_steps": 3882,
4909
+ "num_input_tokens_seen": 32158144,
4910
  "num_train_epochs": 2,
4911
  "save_steps": 50,
4912
  "stateful_callbacks": {
 
4921
  "attributes": {}
4922
  }
4923
  },
4924
+ "total_flos": 2121940811644928.0,
4925
  "train_batch_size": 1,
4926
  "trial_name": null,
4927
  "trial_params": null