ben81828 commited on
Commit
256ae4d
·
verified ·
1 Parent(s): c989684

Training in progress, step 2800, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:dcf51604c7fbe68fb5bec81cac3b4632fb619bd371e32820e3fd872616836861
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2f209712b49d72a0a61d59f7b8f5e51be948aa2ecb02420dc257c2b1ba5b38a8
3
  size 29034840
last-checkpoint/global_step2799/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2046c56a4678d6e319a2f603d0b1e9398b496663d16ed2bf5a036ebfea22add
3
+ size 43429616
last-checkpoint/global_step2799/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:370ccbaecefb9ce961e671a4165e47d7a69af72ff6915ddf34336361fff7160c
3
+ size 43429616
last-checkpoint/global_step2799/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5c5eac2a9b2ebfbebab9b21261d5722ad12e898f434fa74e8730c56b7cb34e8
3
+ size 43429616
last-checkpoint/global_step2799/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15279a29771bcf9661f5dd1bbdfe635434bf014fc851029796f713fc908492ca
3
+ size 43429616
last-checkpoint/global_step2799/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb57e085d33592674e15e103c61ee81ee7e4de7c83d7b3166bbe37cbc7a3f507
3
+ size 637299
last-checkpoint/global_step2799/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d5f662afab50ff642f02642e2e708fff5b1b0e4c174cd936f632115944c55d2
3
+ size 637171
last-checkpoint/global_step2799/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f802cd49f46b34e8f236eac45af3046f74c89f9b8b521ac23f5c2e931c0055ab
3
+ size 637171
last-checkpoint/global_step2799/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86ecedc8766b6b51fc2ce655bc577f9ed891c4aca1eb78f64ebdbdb3bc691604
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step2749
 
1
+ global_step2799
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49db5a9fd0c84d580c671e52905ebeffc155b36537e76ff966d2e82906708999
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff8dba2341c0517760edfde50521977f02a5bd982ffd3bc03de6109439c4f478
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c8ca224562d8d97aaa131b3516288bb99f68d7dcf62170494326662bda0bb206
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d2bf831df9fbade9ac2a8db79798bc2a7b1afb85a78a6e463ec7a7db4acc0f8e
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:56e86a11c89dba78d60e1b2a1855a651b90a5a22ef131ce65d26af83668c154e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8602ff0a0fa366d46b61c0ef2b23ce468387898cf2bc1027e5450de73ddf647f
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dd63019c923e9692431619aced46b91aaf3fd22e1c22ec0a64347f2fe635a0e
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bb51d675cf23603b1b765cd645f53d6b66ddb104d56d48674e9c798e086f696
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d5206e8eee1fd5c926f6f8750c100fc15237f73c48e6ae482a3b1e4fadc9091
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b2248ae5cd08b24968d749ba4f37513ee0a4cfb5059768755ed9ef6b0b1fd810
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.18780523538589478,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-2650",
4
- "epoch": 1.4161730620654134,
5
  "eval_steps": 50,
6
- "global_step": 2750,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -4902,11 +4902,100 @@
4902
  "eval_steps_per_second": 0.769,
4903
  "num_input_tokens_seen": 32158144,
4904
  "step": 2750
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
4905
  }
4906
  ],
4907
  "logging_steps": 5,
4908
  "max_steps": 3882,
4909
- "num_input_tokens_seen": 32158144,
4910
  "num_train_epochs": 2,
4911
  "save_steps": 50,
4912
  "stateful_callbacks": {
@@ -4921,7 +5010,7 @@
4921
  "attributes": {}
4922
  }
4923
  },
4924
- "total_flos": 2121940811644928.0,
4925
  "train_batch_size": 1,
4926
  "trial_name": null,
4927
  "trial_params": null
 
1
  {
2
  "best_metric": 0.18780523538589478,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-2650",
4
+ "epoch": 1.4419263456090652,
5
  "eval_steps": 50,
6
+ "global_step": 2800,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
4902
  "eval_steps_per_second": 0.769,
4903
  "num_input_tokens_seen": 32158144,
4904
  "step": 2750
4905
+ },
4906
+ {
4907
+ "epoch": 1.4187483904197786,
4908
+ "grad_norm": 8.62691802455825,
4909
+ "learning_rate": 2.133572118185313e-05,
4910
+ "loss": 0.2459,
4911
+ "num_input_tokens_seen": 32216656,
4912
+ "step": 2755
4913
+ },
4914
+ {
4915
+ "epoch": 1.4213237187741437,
4916
+ "grad_norm": 11.518284151131217,
4917
+ "learning_rate": 2.1161444173145468e-05,
4918
+ "loss": 0.2656,
4919
+ "num_input_tokens_seen": 32275168,
4920
+ "step": 2760
4921
+ },
4922
+ {
4923
+ "epoch": 1.423899047128509,
4924
+ "grad_norm": 5.717001204241147,
4925
+ "learning_rate": 2.098769060374206e-05,
4926
+ "loss": 0.199,
4927
+ "num_input_tokens_seen": 32333664,
4928
+ "step": 2765
4929
+ },
4930
+ {
4931
+ "epoch": 1.4264743754828741,
4932
+ "grad_norm": 5.060469427631098,
4933
+ "learning_rate": 2.0814463627387744e-05,
4934
+ "loss": 0.2552,
4935
+ "num_input_tokens_seen": 32392088,
4936
+ "step": 2770
4937
+ },
4938
+ {
4939
+ "epoch": 1.4290497038372392,
4940
+ "grad_norm": 10.068742247308663,
4941
+ "learning_rate": 2.0641766388269345e-05,
4942
+ "loss": 0.2123,
4943
+ "num_input_tokens_seen": 32450584,
4944
+ "step": 2775
4945
+ },
4946
+ {
4947
+ "epoch": 1.4316250321916044,
4948
+ "grad_norm": 8.04329511861408,
4949
+ "learning_rate": 2.046960202095866e-05,
4950
+ "loss": 0.1827,
4951
+ "num_input_tokens_seen": 32509096,
4952
+ "step": 2780
4953
+ },
4954
+ {
4955
+ "epoch": 1.4342003605459697,
4956
+ "grad_norm": 5.111975388475009,
4957
+ "learning_rate": 2.0297973650355435e-05,
4958
+ "loss": 0.3038,
4959
+ "num_input_tokens_seen": 32567560,
4960
+ "step": 2785
4961
+ },
4962
+ {
4963
+ "epoch": 1.436775688900335,
4964
+ "grad_norm": 2.2165480876124666,
4965
+ "learning_rate": 2.012688439163075e-05,
4966
+ "loss": 0.2078,
4967
+ "num_input_tokens_seen": 32626056,
4968
+ "step": 2790
4969
+ },
4970
+ {
4971
+ "epoch": 1.4393510172547,
4972
+ "grad_norm": 5.208837069631258,
4973
+ "learning_rate": 1.9956337350170446e-05,
4974
+ "loss": 0.23,
4975
+ "num_input_tokens_seen": 32684520,
4976
+ "step": 2795
4977
+ },
4978
+ {
4979
+ "epoch": 1.4419263456090652,
4980
+ "grad_norm": 14.804471075670579,
4981
+ "learning_rate": 1.978633562151875e-05,
4982
+ "loss": 0.2162,
4983
+ "num_input_tokens_seen": 32743032,
4984
+ "step": 2800
4985
+ },
4986
+ {
4987
+ "epoch": 1.4419263456090652,
4988
+ "eval_loss": 0.2771253287792206,
4989
+ "eval_runtime": 19.4256,
4990
+ "eval_samples_per_second": 3.089,
4991
+ "eval_steps_per_second": 0.772,
4992
+ "num_input_tokens_seen": 32743032,
4993
+ "step": 2800
4994
  }
4995
  ],
4996
  "logging_steps": 5,
4997
  "max_steps": 3882,
4998
+ "num_input_tokens_seen": 32743032,
4999
  "num_train_epochs": 2,
5000
  "save_steps": 50,
5001
  "stateful_callbacks": {
 
5010
  "attributes": {}
5011
  }
5012
  },
5013
+ "total_flos": 2160533754609664.0,
5014
  "train_batch_size": 1,
5015
  "trial_name": null,
5016
  "trial_params": null