masatochi commited on
Commit
cbd5b17
·
verified ·
1 Parent(s): f766bb2

Training in progress, step 80, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4e672f97b29a9edce6f91b2e15c39845acd8e5b4753ab13d398e905d5543c13a
3
  size 59827904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f13ba07bc586558fb64f1b66825209f18db56ce572477ab8155107c3ff3f0a88
3
  size 59827904
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:aea7773082e0f49ff710e5dd6702209f08402f1b7153c857b70854dc0dd63a98
3
  size 30875540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:030da15cbb81bbe486c293ede63079918942591b2c0aab7e2ba0cf09fe49d1a5
3
  size 30875540
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02cc5e0d782e790ab911b5caf87830bdc24d1c04827eeec883676aadfd733f51
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:646ade6023f73d01b10ebe8ac45df7f64238b06f8264b4a748bbde983b0bdd8f
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:52408658f8afed2d0ac64c41a3c3f93b4c8a8478fa6362941012d17f634f7dad
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bc75fc1c14b28d29d31fa9d4252536c919fc25a390fac3a1e8c09d6575b4029
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.03667705849990831,
5
  "eval_steps": 34,
6
- "global_step": 75,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -556,6 +556,41 @@
556
  "learning_rate": 0.00016736956436465573,
557
  "loss": 9.1275,
558
  "step": 75
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
559
  }
560
  ],
561
  "logging_steps": 1,
@@ -575,7 +610,7 @@
575
  "attributes": {}
576
  }
577
  },
578
- "total_flos": 1.653359756967936e+17,
579
  "train_batch_size": 3,
580
  "trial_name": null,
581
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.039122195733235526,
5
  "eval_steps": 34,
6
+ "global_step": 80,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
556
  "learning_rate": 0.00016736956436465573,
557
  "loss": 9.1275,
558
  "step": 75
559
+ },
560
+ {
561
+ "epoch": 0.03716608594657375,
562
+ "grad_norm": Infinity,
563
+ "learning_rate": 0.0001659924534878723,
564
+ "loss": 8.57,
565
+ "step": 76
566
+ },
567
+ {
568
+ "epoch": 0.037655113393239195,
569
+ "grad_norm": Infinity,
570
+ "learning_rate": 0.00016459280624867874,
571
+ "loss": 8.7525,
572
+ "step": 77
573
+ },
574
+ {
575
+ "epoch": 0.03814414083990464,
576
+ "grad_norm": 1.074333141353418e+19,
577
+ "learning_rate": 0.0001631711006253251,
578
+ "loss": 8.9513,
579
+ "step": 78
580
+ },
581
+ {
582
+ "epoch": 0.03863316828657008,
583
+ "grad_norm": Infinity,
584
+ "learning_rate": 0.0001617278221289793,
585
+ "loss": 8.9845,
586
+ "step": 79
587
+ },
588
+ {
589
+ "epoch": 0.039122195733235526,
590
+ "grad_norm": Infinity,
591
+ "learning_rate": 0.00016026346363792567,
592
+ "loss": 9.2195,
593
+ "step": 80
594
  }
595
  ],
596
  "logging_steps": 1,
 
610
  "attributes": {}
611
  }
612
  },
613
+ "total_flos": 1.7635837407657984e+17,
614
  "train_batch_size": 3,
615
  "trial_name": null,
616
  "trial_params": null