ben81828 commited on
Commit
be1b786
·
verified ·
1 Parent(s): 50d7a49

Training in progress, step 900, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2604e713adfd5ff8fff3e76524b74b5484d4e4df2c3ddf9459067c342c4d295
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dca569a5bee748c34c51ac0d5b0f16a410ab4a71a296a7c172247ce0d6801519
3
  size 29034840
last-checkpoint/global_step900/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce9f5392302612002fef358fde0e7f6f4d7930bd85ca39b4ccadf3e35ec527a2
3
+ size 43429616
last-checkpoint/global_step900/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7fd03ef73f13a5f642a2a63469ec5fb1da3f7a708d488660ccb4dc3e8b83555
3
+ size 43429616
last-checkpoint/global_step900/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aee70bb579dfa57fb6ad3b18a5d3cdad2be777077c46fcb523727e225c970391
3
+ size 43429616
last-checkpoint/global_step900/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0eff7aa86298b36602e11144d5d197575c29ab1a5267d726c14a5d6c537bc77f
3
+ size 43429616
last-checkpoint/global_step900/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7162c4e09532f9490922bccb0dbbcfe539972e0f3d49af32428a76030c053274
3
+ size 637299
last-checkpoint/global_step900/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8a1fabfaa7ff216589f1f45e0de64de68782d934b86fc832996ada7c681c3bd9
3
+ size 637171
last-checkpoint/global_step900/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c68864aa3523e8a7232f1228e6e0dba67ac184874fa8332362ced37b3cbe041a
3
+ size 637171
last-checkpoint/global_step900/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:834a420bd57a2a0fe552734d61c0bcc178574ff6e356d461ed6815bc79695c74
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step850
 
1
+ global_step900
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:36c9044354f826de248840acaaec171f816609c147a664089731a0570deef948
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8f22ced19e790cc864cefe3b7c711d9ae631c44f95d42fb4829688cc3de0153
3
  size 15024
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:54f4c4013326db4e7267b656aaf72b86570f8aeee91ad39242a416cf8b963191
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e0407513eba77d34cbf3adf0e59a58bd80716f4f00f414854253637e82be43d
3
  size 15024
last-checkpoint/rng_state_2.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43e910793831957d8685c316138c33eef8867edf60052477dc9ad6ec0c6da901
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6060636c023258ce9b965e244b8a58b4c99d5784dde4405b39737550ef50cd4f
3
  size 15024
last-checkpoint/rng_state_3.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7cfd0d71ef5a6c58d9f1d46851f4b1e699ca8a50ab3223cfb39668895cffeef2
3
  size 15024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c24ccdfdcde39cb2265c82c50c36ffdfcc670f757aba4bcf4bb0fdc6d1373c4c
3
  size 15024
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e63e21db765af1a20657ecdebe434ff7baf8429eec4fb44a9d43fe347139b971
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:728e48e0d034d64ded7c746a218c0b746489e6573b6bc352ba39a730bda96bbd
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.7039459347724915,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-850",
4
- "epoch": 0.43780582024208087,
5
  "eval_steps": 50,
6
- "global_step": 850,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1520,11 +1520,100 @@
1520
  "eval_steps_per_second": 0.758,
1521
  "num_input_tokens_seen": 9941896,
1522
  "step": 850
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1523
  }
1524
  ],
1525
  "logging_steps": 5,
1526
  "max_steps": 3882,
1527
- "num_input_tokens_seen": 9941896,
1528
  "num_train_epochs": 2,
1529
  "save_steps": 50,
1530
  "stateful_callbacks": {
@@ -1539,7 +1628,7 @@
1539
  "attributes": {}
1540
  }
1541
  },
1542
- "total_flos": 655954683887616.0,
1543
  "train_batch_size": 1,
1544
  "trial_name": null,
1545
  "trial_params": null
 
1
  {
2
  "best_metric": 0.7039459347724915,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_stenosis_classily_scale4/lora/sft/checkpoint-850",
4
+ "epoch": 0.4635591037857327,
5
  "eval_steps": 50,
6
+ "global_step": 900,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1520
  "eval_steps_per_second": 0.758,
1521
  "num_input_tokens_seen": 9941896,
1522
  "step": 850
1523
+ },
1524
+ {
1525
+ "epoch": 0.44038114859644606,
1526
+ "grad_norm": 8.584709523479143,
1527
+ "learning_rate": 9.22997419657528e-05,
1528
+ "loss": 0.6829,
1529
+ "num_input_tokens_seen": 10000336,
1530
+ "step": 855
1531
+ },
1532
+ {
1533
+ "epoch": 0.44295647695081125,
1534
+ "grad_norm": 9.894526327436065,
1535
+ "learning_rate": 9.218577901126713e-05,
1536
+ "loss": 0.6971,
1537
+ "num_input_tokens_seen": 10058816,
1538
+ "step": 860
1539
+ },
1540
+ {
1541
+ "epoch": 0.4455318053051764,
1542
+ "grad_norm": 6.844807331677797,
1543
+ "learning_rate": 9.207105035634397e-05,
1544
+ "loss": 0.7239,
1545
+ "num_input_tokens_seen": 10117320,
1546
+ "step": 865
1547
+ },
1548
+ {
1549
+ "epoch": 0.4481071336595416,
1550
+ "grad_norm": 9.911971184322747,
1551
+ "learning_rate": 9.195555808338603e-05,
1552
+ "loss": 0.7113,
1553
+ "num_input_tokens_seen": 10175824,
1554
+ "step": 870
1555
+ },
1556
+ {
1557
+ "epoch": 0.45068246201390677,
1558
+ "grad_norm": 11.42729729117445,
1559
+ "learning_rate": 9.183930428865622e-05,
1560
+ "loss": 0.685,
1561
+ "num_input_tokens_seen": 10234288,
1562
+ "step": 875
1563
+ },
1564
+ {
1565
+ "epoch": 0.45325779036827196,
1566
+ "grad_norm": 6.6164653318829005,
1567
+ "learning_rate": 9.17222910822396e-05,
1568
+ "loss": 0.6804,
1569
+ "num_input_tokens_seen": 10292736,
1570
+ "step": 880
1571
+ },
1572
+ {
1573
+ "epoch": 0.45583311872263715,
1574
+ "grad_norm": 9.566984327413312,
1575
+ "learning_rate": 9.160452058800504e-05,
1576
+ "loss": 0.7056,
1577
+ "num_input_tokens_seen": 10351224,
1578
+ "step": 885
1579
+ },
1580
+ {
1581
+ "epoch": 0.4584084470770023,
1582
+ "grad_norm": 8.37543764212447,
1583
+ "learning_rate": 9.148599494356671e-05,
1584
+ "loss": 0.7234,
1585
+ "num_input_tokens_seen": 10409736,
1586
+ "step": 890
1587
+ },
1588
+ {
1589
+ "epoch": 0.4609837754313675,
1590
+ "grad_norm": 8.118764978921215,
1591
+ "learning_rate": 9.136671630024527e-05,
1592
+ "loss": 0.7505,
1593
+ "num_input_tokens_seen": 10468240,
1594
+ "step": 895
1595
+ },
1596
+ {
1597
+ "epoch": 0.4635591037857327,
1598
+ "grad_norm": 8.91603360160678,
1599
+ "learning_rate": 9.124668682302882e-05,
1600
+ "loss": 0.7378,
1601
+ "num_input_tokens_seen": 10526712,
1602
+ "step": 900
1603
+ },
1604
+ {
1605
+ "epoch": 0.4635591037857327,
1606
+ "eval_loss": 0.7197856903076172,
1607
+ "eval_runtime": 19.6818,
1608
+ "eval_samples_per_second": 3.048,
1609
+ "eval_steps_per_second": 0.762,
1610
+ "num_input_tokens_seen": 10526712,
1611
+ "step": 900
1612
  }
1613
  ],
1614
  "logging_steps": 5,
1615
  "max_steps": 3882,
1616
+ "num_input_tokens_seen": 10526712,
1617
  "num_train_epochs": 2,
1618
  "save_steps": 50,
1619
  "stateful_callbacks": {
 
1628
  "attributes": {}
1629
  }
1630
  },
1631
+ "total_flos": 694543977283584.0,
1632
  "train_batch_size": 1,
1633
  "trial_name": null,
1634
  "trial_params": null