masatochi commited on
Commit
ea6ce40
·
verified ·
1 Parent(s): 4e49e8b

Training in progress, step 10, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fcd66902054a93e8e0d714260f2f2075546de0d24e2a40f74a9cc256211f182a
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9ea66dd4489f55c292edd6ab8d9352d70360f9ad5a4ceab786a717ee5fd7974
3
  size 83945296
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8c463401ffa91e44167009cab353d2f77a0606563274e968235352e5ed588f26
3
  size 43122580
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aeca6749dd845c34cfa3cdf59a1bc387ddede928591cc424678f55715fb2f2f
3
  size 43122580
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5d11172069d3809e4b1d435d13824f7b7bfef9c55e762daff1ad8efa9119dcca
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1bf24cd4bf25e308928bff2034c8266e5fe111aaec9b6d5c14e665ac843e0bc
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16178af669793c21eac8710459f1cc10bcf09af1984c5f2d80e76b26b329c8ba
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67680b8b35ea230d53acee387012f8809825a2c511bd41340f42320f849e349d
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.0024451372333272204,
5
  "eval_steps": 34,
6
- "global_step": 5,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -50,6 +50,41 @@
50
  "learning_rate": 3.3333333333333335e-05,
51
  "loss": 1.1686,
52
  "step": 5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  }
54
  ],
55
  "logging_steps": 1,
@@ -69,7 +104,7 @@
69
  "attributes": {}
70
  }
71
  },
72
- "total_flos": 2.219477097775104e+16,
73
  "train_batch_size": 3,
74
  "trial_name": null,
75
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.004890274466654441,
5
  "eval_steps": 34,
6
+ "global_step": 10,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
50
  "learning_rate": 3.3333333333333335e-05,
51
  "loss": 1.1686,
52
  "step": 5
53
+ },
54
+ {
55
+ "epoch": 0.0029341646799926646,
56
+ "grad_norm": 0.44767728447914124,
57
+ "learning_rate": 4e-05,
58
+ "loss": 1.2578,
59
+ "step": 6
60
+ },
61
+ {
62
+ "epoch": 0.0034231921266581087,
63
+ "grad_norm": 0.4529785215854645,
64
+ "learning_rate": 4.666666666666667e-05,
65
+ "loss": 1.2452,
66
+ "step": 7
67
+ },
68
+ {
69
+ "epoch": 0.0039122195733235525,
70
+ "grad_norm": 0.6146576404571533,
71
+ "learning_rate": 5.333333333333333e-05,
72
+ "loss": 1.5423,
73
+ "step": 8
74
+ },
75
+ {
76
+ "epoch": 0.004401247019988997,
77
+ "grad_norm": 0.42456191778182983,
78
+ "learning_rate": 6e-05,
79
+ "loss": 1.1323,
80
+ "step": 9
81
+ },
82
+ {
83
+ "epoch": 0.004890274466654441,
84
+ "grad_norm": 0.4658707082271576,
85
+ "learning_rate": 6.666666666666667e-05,
86
+ "loss": 1.2595,
87
+ "step": 10
88
  }
89
  ],
90
  "logging_steps": 1,
 
104
  "attributes": {}
105
  }
106
  },
107
+ "total_flos": 4.438954195550208e+16,
108
  "train_batch_size": 3,
109
  "trial_name": null,
110
  "trial_params": null