ben81828 commited on
Commit
4a55b96
·
verified ·
1 Parent(s): 0263caa

Training in progress, step 1050, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eec35e19707d55e085c42f2fc7f96384949ae1c2b6909dea425372d9568c8a7a
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ec2597cb66cc8bf7c12cf0511653b2a0fba5900300120225835d32ef88d42ca
3
  size 29034840
last-checkpoint/global_step1050/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de2e167999736a3e47f82ccd89dd02140ccd2fb9e1416493e0d58739b4df6c86
3
+ size 43429616
last-checkpoint/global_step1050/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65ffb0f2b0651eaf6c98bc5de8e998c605ae188e16936b9958297b9809eb66a7
3
+ size 43429616
last-checkpoint/global_step1050/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94cf56191a583e6143daf3d907055192e9d16a3a366fe400ec28191e559cadb2
3
+ size 43429616
last-checkpoint/global_step1050/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ccfaaa1c7bf29b7d738cd613b7a8c1cc7a2aad384ef1de5c5ce601c568d6969
3
+ size 43429616
last-checkpoint/global_step1050/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5740907eeaf461a960ab6da055d43d89f41f6417899be566427552f76b967abe
3
+ size 637299
last-checkpoint/global_step1050/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28dcc960a14c72ef47fb9722c4b98cf1679b58db19513a1020156883879227c7
3
+ size 637171
last-checkpoint/global_step1050/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:055280373e3009fe56853d9b9ef4c853cc09e485fa66a9bd20cba222d17c0bf5
3
+ size 637171
last-checkpoint/global_step1050/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:179e95a89c200a847bd6de0c47ca8b3d721ccf69a889371f83c8cb19695431d6
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step1000
 
1
+ global_step1050
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7d73dfcc09cf3d6f08149535e03920234febc15f7e9a166987f3bc01ee871abf
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67db742b8aa1744a8224bf2a1f79d89caff63b15f78a455d92bb666df82183ea
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4396a64b6da4868d060d1e3c7c9ccb12c39d63bd0f7b146d2512400aff4c769c
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0f9acb7e6f8bbfb305c3601c71eb6189af24942fab5f99046412c03bb10c3eb
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95877efc8fb5eb302819ee7effca4222569cdcfdebb9fa5d9846e68ed9e833fe
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:713783338342f7486f6f186abd03c5963a0d22368f403efb2bf903ed083d2b64
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9fa4f23377f00fdde731da68a8690098617a1fdd912e03cdaa8bde87c493179
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:83aaec0b5b7d8a2da4577075066cf434ce6e9feb9327edbea6677a2e51d76466
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:22246d81041447d5284c93bc1d01007e3c5bfb83233c72fb1f7878544bb0fc38
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:15418e5c97d5f142900d492dc3d977b1b3706e9b18920657077086449e4a88dc
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "best_metric": 0.6830747723579407,
3
- "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-950",
4
- "epoch": 0.5150656708730363,
5
  "eval_steps": 50,
6
- "global_step": 1000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1787,11 +1787,100 @@
1787
  "eval_steps_per_second": 0.762,
1788
  "num_input_tokens_seen": 11696200,
1789
  "step": 1000
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1790
  }
1791
  ],
1792
  "logging_steps": 5,
1793
  "max_steps": 3882,
1794
- "num_input_tokens_seen": 11696200,
1795
  "num_train_epochs": 2,
1796
  "save_steps": 50,
1797
  "stateful_callbacks": {
@@ -1806,7 +1895,7 @@
1806
  "attributes": {}
1807
  }
1808
  },
1809
- "total_flos": 771713061093376.0,
1810
  "train_batch_size": 1,
1811
  "trial_name": null,
1812
  "trial_params": null
 
1
  {
2
+ "best_metric": 0.6824291348457336,
3
+ "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-1050",
4
+ "epoch": 0.5408189544166881,
5
  "eval_steps": 50,
6
+ "global_step": 1050,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1787
  "eval_steps_per_second": 0.762,
1788
  "num_input_tokens_seen": 11696200,
1789
  "step": 1000
1790
+ },
1791
+ {
1792
+ "epoch": 0.5176409992274015,
1793
+ "grad_norm": 9.38238507668501,
1794
+ "learning_rate": 8.855660246460895e-05,
1795
+ "loss": 0.6959,
1796
+ "num_input_tokens_seen": 11754720,
1797
+ "step": 1005
1798
+ },
1799
+ {
1800
+ "epoch": 0.5202163275817667,
1801
+ "grad_norm": 8.602013574457626,
1802
+ "learning_rate": 8.842062953648023e-05,
1803
+ "loss": 0.6918,
1804
+ "num_input_tokens_seen": 11813216,
1805
+ "step": 1010
1806
+ },
1807
+ {
1808
+ "epoch": 0.5227916559361319,
1809
+ "grad_norm": 6.752768596735988,
1810
+ "learning_rate": 8.828395924792497e-05,
1811
+ "loss": 0.7269,
1812
+ "num_input_tokens_seen": 11871712,
1813
+ "step": 1015
1814
+ },
1815
+ {
1816
+ "epoch": 0.525366984290497,
1817
+ "grad_norm": 3.9336242665009187,
1818
+ "learning_rate": 8.814659407960141e-05,
1819
+ "loss": 0.7026,
1820
+ "num_input_tokens_seen": 11930200,
1821
+ "step": 1020
1822
+ },
1823
+ {
1824
+ "epoch": 0.5279423126448622,
1825
+ "grad_norm": 5.739197643968765,
1826
+ "learning_rate": 8.800853652478028e-05,
1827
+ "loss": 0.6467,
1828
+ "num_input_tokens_seen": 11988704,
1829
+ "step": 1025
1830
+ },
1831
+ {
1832
+ "epoch": 0.5305176409992274,
1833
+ "grad_norm": 4.583970239100745,
1834
+ "learning_rate": 8.786978908929966e-05,
1835
+ "loss": 0.6155,
1836
+ "num_input_tokens_seen": 12047176,
1837
+ "step": 1030
1838
+ },
1839
+ {
1840
+ "epoch": 0.5330929693535926,
1841
+ "grad_norm": 7.891614334520996,
1842
+ "learning_rate": 8.773035429151937e-05,
1843
+ "loss": 0.7849,
1844
+ "num_input_tokens_seen": 12105680,
1845
+ "step": 1035
1846
+ },
1847
+ {
1848
+ "epoch": 0.5356682977079578,
1849
+ "grad_norm": 10.308514200015722,
1850
+ "learning_rate": 8.759023466227538e-05,
1851
+ "loss": 0.6341,
1852
+ "num_input_tokens_seen": 12164208,
1853
+ "step": 1040
1854
+ },
1855
+ {
1856
+ "epoch": 0.5382436260623229,
1857
+ "grad_norm": 5.0434906291132995,
1858
+ "learning_rate": 8.744943274483376e-05,
1859
+ "loss": 0.7189,
1860
+ "num_input_tokens_seen": 12222672,
1861
+ "step": 1045
1862
+ },
1863
+ {
1864
+ "epoch": 0.5408189544166881,
1865
+ "grad_norm": 4.3092898145567125,
1866
+ "learning_rate": 8.730795109484461e-05,
1867
+ "loss": 0.6914,
1868
+ "num_input_tokens_seen": 12281072,
1869
+ "step": 1050
1870
+ },
1871
+ {
1872
+ "epoch": 0.5408189544166881,
1873
+ "eval_loss": 0.6824291348457336,
1874
+ "eval_runtime": 19.7949,
1875
+ "eval_samples_per_second": 3.031,
1876
+ "eval_steps_per_second": 0.758,
1877
+ "num_input_tokens_seen": 12281072,
1878
+ "step": 1050
1879
  }
1880
  ],
1881
  "logging_steps": 5,
1882
  "max_steps": 3882,
1883
+ "num_input_tokens_seen": 12281072,
1884
  "num_train_epochs": 2,
1885
  "save_steps": 50,
1886
  "stateful_callbacks": {
 
1895
  "attributes": {}
1896
  }
1897
  },
1898
+ "total_flos": 810306416410624.0,
1899
  "train_batch_size": 1,
1900
  "trial_name": null,
1901
  "trial_params": null