angelitasr
commited on
End of training
Browse files
README.md
CHANGED
@@ -236,7 +236,7 @@ You can finetune this model on your own dataset.
|
|
236 |
### Training Hyperparameters
|
237 |
#### Non-Default Hyperparameters
|
238 |
|
239 |
-
- `num_train_epochs`:
|
240 |
- `fp16`: True
|
241 |
- `push_to_hub`: True
|
242 |
- `batch_sampler`: no_duplicates
|
@@ -261,7 +261,7 @@ You can finetune this model on your own dataset.
|
|
261 |
- `adam_beta2`: 0.999
|
262 |
- `adam_epsilon`: 1e-08
|
263 |
- `max_grad_norm`: 1.0
|
264 |
-
- `num_train_epochs`:
|
265 |
- `max_steps`: -1
|
266 |
- `lr_scheduler_type`: linear
|
267 |
- `lr_scheduler_kwargs`: {}
|
@@ -362,10 +362,14 @@ You can finetune this model on your own dataset.
|
|
362 |
### Training Logs
|
363 |
| Epoch | Step | Training Loss |
|
364 |
|:------:|:----:|:-------------:|
|
365 |
-
| 1.1416 | 500 | 0.
|
366 |
-
| 2.2831 | 1000 | 0.
|
367 |
-
| 3.4247 | 1500 | 0.
|
368 |
-
| 4.5662 | 2000 | 0.
|
|
|
|
|
|
|
|
|
369 |
|
370 |
|
371 |
### Framework Versions
|
|
|
236 |
### Training Hyperparameters
|
237 |
#### Non-Default Hyperparameters
|
238 |
|
239 |
+
- `num_train_epochs`: 10
|
240 |
- `fp16`: True
|
241 |
- `push_to_hub`: True
|
242 |
- `batch_sampler`: no_duplicates
|
|
|
261 |
- `adam_beta2`: 0.999
|
262 |
- `adam_epsilon`: 1e-08
|
263 |
- `max_grad_norm`: 1.0
|
264 |
+
- `num_train_epochs`: 10
|
265 |
- `max_steps`: -1
|
266 |
- `lr_scheduler_type`: linear
|
267 |
- `lr_scheduler_kwargs`: {}
|
|
|
362 |
### Training Logs
|
363 |
| Epoch | Step | Training Loss |
|
364 |
|:------:|:----:|:-------------:|
|
365 |
+
| 1.1416 | 500 | 0.341 |
|
366 |
+
| 2.2831 | 1000 | 0.1082 |
|
367 |
+
| 3.4247 | 1500 | 0.0485 |
|
368 |
+
| 4.5662 | 2000 | 0.0226 |
|
369 |
+
| 5.7078 | 2500 | 0.0133 |
|
370 |
+
| 6.8493 | 3000 | 0.0066 |
|
371 |
+
| 7.9909 | 3500 | 0.0042 |
|
372 |
+
| 9.1324 | 4000 | 0.0017 |
|
373 |
|
374 |
|
375 |
### Framework Versions
|