angelitasr commited on
Commit
df6d12e
·
verified ·
1 Parent(s): 5c2cf9a

End of training

Browse files
Files changed (1) hide show
  1. README.md +10 -6
README.md CHANGED
@@ -236,7 +236,7 @@ You can finetune this model on your own dataset.
236
  ### Training Hyperparameters
237
  #### Non-Default Hyperparameters
238
 
239
- - `num_train_epochs`: 5
240
  - `fp16`: True
241
  - `push_to_hub`: True
242
  - `batch_sampler`: no_duplicates
@@ -261,7 +261,7 @@ You can finetune this model on your own dataset.
261
  - `adam_beta2`: 0.999
262
  - `adam_epsilon`: 1e-08
263
  - `max_grad_norm`: 1.0
264
- - `num_train_epochs`: 5
265
  - `max_steps`: -1
266
  - `lr_scheduler_type`: linear
267
  - `lr_scheduler_kwargs`: {}
@@ -362,10 +362,14 @@ You can finetune this model on your own dataset.
362
  ### Training Logs
363
  | Epoch | Step | Training Loss |
364
  |:------:|:----:|:-------------:|
365
- | 1.1416 | 500 | 0.3382 |
366
- | 2.2831 | 1000 | 0.1004 |
367
- | 3.4247 | 1500 | 0.0386 |
368
- | 4.5662 | 2000 | 0.0133 |
 
 
 
 
369
 
370
 
371
  ### Framework Versions
 
236
  ### Training Hyperparameters
237
  #### Non-Default Hyperparameters
238
 
239
+ - `num_train_epochs`: 10
240
  - `fp16`: True
241
  - `push_to_hub`: True
242
  - `batch_sampler`: no_duplicates
 
261
  - `adam_beta2`: 0.999
262
  - `adam_epsilon`: 1e-08
263
  - `max_grad_norm`: 1.0
264
+ - `num_train_epochs`: 10
265
  - `max_steps`: -1
266
  - `lr_scheduler_type`: linear
267
  - `lr_scheduler_kwargs`: {}
 
362
  ### Training Logs
363
  | Epoch | Step | Training Loss |
364
  |:------:|:----:|:-------------:|
365
+ | 1.1416 | 500 | 0.341 |
366
+ | 2.2831 | 1000 | 0.1082 |
367
+ | 3.4247 | 1500 | 0.0485 |
368
+ | 4.5662 | 2000 | 0.0226 |
369
+ | 5.7078 | 2500 | 0.0133 |
370
+ | 6.8493 | 3000 | 0.0066 |
371
+ | 7.9909 | 3500 | 0.0042 |
372
+ | 9.1324 | 4000 | 0.0017 |
373
 
374
 
375
  ### Framework Versions