fats-fme commited on
Commit
80c8a62
·
verified ·
1 Parent(s): 2b0a744

Training in progress, step 45, checkpoint

Browse files
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af9e788e546f18a3b27058939ea4434cdddac18c862a33dfd4b9a0f72f9921d8
3
  size 239452242
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3b0162c125dace0461332123e8c7d602462e84e158ad2d3f071bbc125b34b8d
3
  size 239452242
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9beca6c32763a3cc8051aec28baf91f153cdbe56a51f49590523253e9deac3c0
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:56e0b4692f0d79aa027250daa798b9b99aa2daab99fce81150537a9623556468
3
  size 14512
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b1384fee391a6ad2405d499f49efeb5a52b5f8c74d45b3ac0747ecb863c12d32
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a7c19a532ddf743881461eea886de51a0451550ba062c994babb62565733b22d
3
  size 14512
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3fbc936ac0284bc0c88ce620e190b6b6984693d58131c42013c320d4b2c3a45e
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e94f4caf902c120d48372da3bb1cf8d147f3ddef7262ea875dbd118eb698ece
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.4979253112033195,
5
  "eval_steps": 15,
6
- "global_step": 30,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -241,6 +241,119 @@
241
  "eval_samples_per_second": 6.084,
242
  "eval_steps_per_second": 1.551,
243
  "step": 30
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
244
  }
245
  ],
246
  "logging_steps": 1,
@@ -260,7 +373,7 @@
260
  "attributes": {}
261
  }
262
  },
263
- "total_flos": 8.85317179539456e+16,
264
  "train_batch_size": 2,
265
  "trial_name": null,
266
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.7468879668049793,
5
  "eval_steps": 15,
6
+ "global_step": 45,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
241
  "eval_samples_per_second": 6.084,
242
  "eval_steps_per_second": 1.551,
243
  "step": 30
244
+ },
245
+ {
246
+ "epoch": 0.5145228215767634,
247
+ "grad_norm": NaN,
248
+ "learning_rate": 0.00016494480483301836,
249
+ "loss": 0.0,
250
+ "step": 31
251
+ },
252
+ {
253
+ "epoch": 0.5311203319502075,
254
+ "grad_norm": NaN,
255
+ "learning_rate": 0.00015877852522924732,
256
+ "loss": 0.0,
257
+ "step": 32
258
+ },
259
+ {
260
+ "epoch": 0.5477178423236515,
261
+ "grad_norm": NaN,
262
+ "learning_rate": 0.0001522498564715949,
263
+ "loss": 0.0,
264
+ "step": 33
265
+ },
266
+ {
267
+ "epoch": 0.5643153526970954,
268
+ "grad_norm": NaN,
269
+ "learning_rate": 0.00014539904997395468,
270
+ "loss": 0.0,
271
+ "step": 34
272
+ },
273
+ {
274
+ "epoch": 0.5809128630705395,
275
+ "grad_norm": NaN,
276
+ "learning_rate": 0.000138268343236509,
277
+ "loss": 0.0,
278
+ "step": 35
279
+ },
280
+ {
281
+ "epoch": 0.5975103734439834,
282
+ "grad_norm": NaN,
283
+ "learning_rate": 0.00013090169943749476,
284
+ "loss": 0.0,
285
+ "step": 36
286
+ },
287
+ {
288
+ "epoch": 0.6141078838174274,
289
+ "grad_norm": NaN,
290
+ "learning_rate": 0.00012334453638559057,
291
+ "loss": 0.0,
292
+ "step": 37
293
+ },
294
+ {
295
+ "epoch": 0.6307053941908713,
296
+ "grad_norm": NaN,
297
+ "learning_rate": 0.0001156434465040231,
298
+ "loss": 0.0,
299
+ "step": 38
300
+ },
301
+ {
302
+ "epoch": 0.6473029045643154,
303
+ "grad_norm": NaN,
304
+ "learning_rate": 0.0001078459095727845,
305
+ "loss": 0.0,
306
+ "step": 39
307
+ },
308
+ {
309
+ "epoch": 0.6639004149377593,
310
+ "grad_norm": NaN,
311
+ "learning_rate": 0.0001,
312
+ "loss": 0.0,
313
+ "step": 40
314
+ },
315
+ {
316
+ "epoch": 0.6804979253112033,
317
+ "grad_norm": NaN,
318
+ "learning_rate": 9.215409042721552e-05,
319
+ "loss": 0.0,
320
+ "step": 41
321
+ },
322
+ {
323
+ "epoch": 0.6970954356846473,
324
+ "grad_norm": NaN,
325
+ "learning_rate": 8.435655349597689e-05,
326
+ "loss": 0.0,
327
+ "step": 42
328
+ },
329
+ {
330
+ "epoch": 0.7136929460580913,
331
+ "grad_norm": NaN,
332
+ "learning_rate": 7.66554636144095e-05,
333
+ "loss": 0.0,
334
+ "step": 43
335
+ },
336
+ {
337
+ "epoch": 0.7302904564315352,
338
+ "grad_norm": NaN,
339
+ "learning_rate": 6.909830056250527e-05,
340
+ "loss": 0.0,
341
+ "step": 44
342
+ },
343
+ {
344
+ "epoch": 0.7468879668049793,
345
+ "grad_norm": NaN,
346
+ "learning_rate": 6.173165676349103e-05,
347
+ "loss": 0.0,
348
+ "step": 45
349
+ },
350
+ {
351
+ "epoch": 0.7468879668049793,
352
+ "eval_loss": NaN,
353
+ "eval_runtime": 16.8908,
354
+ "eval_samples_per_second": 6.039,
355
+ "eval_steps_per_second": 1.539,
356
+ "step": 45
357
  }
358
  ],
359
  "logging_steps": 1,
 
373
  "attributes": {}
374
  }
375
  },
376
+ "total_flos": 1.327975769309184e+17,
377
  "train_batch_size": 2,
378
  "trial_name": null,
379
  "trial_params": null