akhilfau commited on
Commit
d59a8f4
·
verified ·
1 Parent(s): 032fead

End of training

Browse files
Files changed (2) hide show
  1. README.md +18 -14
  2. generation_config.json +1 -1
README.md CHANGED
@@ -1,8 +1,11 @@
1
  ---
 
2
  license: apache-2.0
3
- base_model: HuggingFaceTB/SmolLM2-135M
4
  tags:
5
  - generated_from_trainer
 
 
6
  model-index:
7
  - name: fine-tuned-smolLM2-physics_10
8
  results: []
@@ -13,9 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # fine-tuned-smolLM2-physics_10
15
 
16
- This model is a fine-tuned version of [HuggingFaceTB/SmolLM2-135M](https://huggingface.co/HuggingFaceTB/SmolLM2-135M) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 0.9803
 
19
 
20
  ## Model description
21
 
@@ -34,26 +38,26 @@ More information needed
34
  ### Training hyperparameters
35
 
36
  The following hyperparameters were used during training:
37
- - learning_rate: 5e-05
38
  - train_batch_size: 2
39
- - eval_batch_size: 8
40
  - seed: 42
41
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
  - num_epochs: 3
44
 
45
  ### Training results
46
 
47
- | Training Loss | Epoch | Step | Validation Loss |
48
- |:-------------:|:-----:|:----:|:---------------:|
49
- | No log | 1.0 | 4 | 1.0561 |
50
- | No log | 2.0 | 8 | 0.9998 |
51
- | 0.8859 | 3.0 | 12 | 0.9803 |
52
 
53
 
54
  ### Framework versions
55
 
56
- - Transformers 4.39.3
57
- - Pytorch 2.3.1
58
  - Datasets 3.1.0
59
- - Tokenizers 0.15.2
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
+ base_model: HuggingFaceTB/SmolLM2-360M
5
  tags:
6
  - generated_from_trainer
7
+ metrics:
8
+ - accuracy
9
  model-index:
10
  - name: fine-tuned-smolLM2-physics_10
11
  results: []
 
16
 
17
  # fine-tuned-smolLM2-physics_10
18
 
19
+ This model is a fine-tuned version of [HuggingFaceTB/SmolLM2-360M](https://huggingface.co/HuggingFaceTB/SmolLM2-360M) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.4107
22
+ - Accuracy: 0.1565
23
 
24
  ## Model description
25
 
 
38
  ### Training hyperparameters
39
 
40
  The following hyperparameters were used during training:
41
+ - learning_rate: 2e-05
42
  - train_batch_size: 2
43
+ - eval_batch_size: 2
44
  - seed: 42
45
+ - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
46
  - lr_scheduler_type: linear
47
  - num_epochs: 3
48
 
49
  ### Training results
50
 
51
+ | Training Loss | Epoch | Step | Validation Loss | Accuracy |
52
+ |:-------------:|:-----:|:----:|:---------------:|:--------:|
53
+ | 0.6273 | 1.0 | 40 | 0.8006 | 0.0231 |
54
+ | 0.3887 | 2.0 | 80 | 0.4202 | 0.1566 |
55
+ | 0.324 | 3.0 | 120 | 0.4107 | 0.1565 |
56
 
57
 
58
  ### Framework versions
59
 
60
+ - Transformers 4.46.2
61
+ - Pytorch 2.4.1+cu121
62
  - Datasets 3.1.0
63
+ - Tokenizers 0.20.3
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
- "transformers_version": "4.39.3"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
  "eos_token_id": 0,
5
+ "transformers_version": "4.46.2"
6
  }