Adokato commited on
Commit
a7e5120
·
verified ·
1 Parent(s): 6f6d6a8

End of training

Browse files
Files changed (5) hide show
  1. README.md +16 -18
  2. config.json +12 -10
  3. generation_config.json +2 -1
  4. model.safetensors +2 -2
  5. training_args.bin +2 -2
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: apache-2.0
3
- base_model: eugenesiow/bart-paraphrase
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,9 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # results
15
 
16
- This model is a fine-tuned version of [eugenesiow/bart-paraphrase](https://huggingface.co/eugenesiow/bart-paraphrase) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Loss: 2.3697
19
 
20
  ## Model description
21
 
@@ -35,30 +35,28 @@ More information needed
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 2e-05
38
- - train_batch_size: 8
39
- - eval_batch_size: 8
40
  - seed: 42
41
  - gradient_accumulation_steps: 2
42
- - total_train_batch_size: 16
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - lr_scheduler_warmup_steps: 500
46
- - num_epochs: 5
47
 
48
  ### Training results
49
 
50
- | Training Loss | Epoch | Step | Validation Loss |
51
- |:-------------:|:-----:|:----:|:---------------:|
52
- | 5.3656 | 0.99 | 93 | 5.0322 |
53
- | 4.52 | 2.0 | 187 | 3.9055 |
54
- | 3.7139 | 2.99 | 280 | 3.1498 |
55
- | 3.348 | 4.0 | 374 | 2.6742 |
56
- | 2.9365 | 4.97 | 465 | 2.3143 |
57
 
58
 
59
  ### Framework versions
60
 
61
- - Transformers 4.39.3
62
- - Pytorch 2.1.2
63
- - Datasets 2.18.0
64
- - Tokenizers 0.15.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: facebook/bart-base
4
  tags:
5
  - generated_from_trainer
6
  model-index:
 
13
 
14
  # results
15
 
16
+ This model is a fine-tuned version of [facebook/bart-base](https://huggingface.co/facebook/bart-base) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 1.4561
19
 
20
  ## Model description
21
 
 
35
 
36
  The following hyperparameters were used during training:
37
  - learning_rate: 2e-05
38
+ - train_batch_size: 4
39
+ - eval_batch_size: 4
40
  - seed: 42
41
  - gradient_accumulation_steps: 2
42
+ - total_train_batch_size: 8
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
  - lr_scheduler_warmup_steps: 500
46
+ - num_epochs: 3
47
 
48
  ### Training results
49
 
50
+ | Training Loss | Epoch | Step | Validation Loss |
51
+ |:-------------:|:------:|:----:|:---------------:|
52
+ | 3.0606 | 0.9993 | 747 | 2.3514 |
53
+ | 1.9003 | 2.0 | 1495 | 1.7026 |
54
+ | 1.8536 | 2.9980 | 2241 | 1.4560 |
 
 
55
 
56
 
57
  ### Framework versions
58
 
59
+ - Transformers 4.40.0
60
+ - Pytorch 2.2.1+cu121
61
+ - Datasets 2.19.0
62
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "eugenesiow/bart-paraphrase",
3
  "activation_dropout": 0.1,
4
  "activation_function": "gelu",
5
  "add_bias_logits": false,
@@ -11,19 +11,20 @@
11
  "bos_token_id": 0,
12
  "classif_dropout": 0.1,
13
  "classifier_dropout": 0.0,
14
- "d_model": 1024,
15
- "decoder_attention_heads": 16,
16
- "decoder_ffn_dim": 4096,
17
  "decoder_layerdrop": 0.0,
18
- "decoder_layers": 12,
19
  "decoder_start_token_id": 2,
20
  "dropout": 0.1,
21
  "early_stopping": true,
22
- "encoder_attention_heads": 16,
23
- "encoder_ffn_dim": 4096,
24
  "encoder_layerdrop": 0.0,
25
- "encoder_layers": 12,
26
  "eos_token_id": 2,
 
27
  "forced_eos_token_id": 2,
28
  "gradient_checkpointing": false,
29
  "id2label": {
@@ -42,8 +43,9 @@
42
  "model_type": "bart",
43
  "no_repeat_ngram_size": 3,
44
  "normalize_before": false,
 
45
  "num_beams": 4,
46
- "num_hidden_layers": 12,
47
  "pad_token_id": 1,
48
  "scale_embedding": false,
49
  "task_specific_params": {
@@ -67,7 +69,7 @@
67
  }
68
  },
69
  "torch_dtype": "float32",
70
- "transformers_version": "4.39.3",
71
  "use_cache": true,
72
  "vocab_size": 50265
73
  }
 
1
  {
2
+ "_name_or_path": "facebook/bart-base",
3
  "activation_dropout": 0.1,
4
  "activation_function": "gelu",
5
  "add_bias_logits": false,
 
11
  "bos_token_id": 0,
12
  "classif_dropout": 0.1,
13
  "classifier_dropout": 0.0,
14
+ "d_model": 768,
15
+ "decoder_attention_heads": 12,
16
+ "decoder_ffn_dim": 3072,
17
  "decoder_layerdrop": 0.0,
18
+ "decoder_layers": 6,
19
  "decoder_start_token_id": 2,
20
  "dropout": 0.1,
21
  "early_stopping": true,
22
+ "encoder_attention_heads": 12,
23
+ "encoder_ffn_dim": 3072,
24
  "encoder_layerdrop": 0.0,
25
+ "encoder_layers": 6,
26
  "eos_token_id": 2,
27
+ "forced_bos_token_id": 0,
28
  "forced_eos_token_id": 2,
29
  "gradient_checkpointing": false,
30
  "id2label": {
 
43
  "model_type": "bart",
44
  "no_repeat_ngram_size": 3,
45
  "normalize_before": false,
46
+ "normalize_embedding": true,
47
  "num_beams": 4,
48
+ "num_hidden_layers": 6,
49
  "pad_token_id": 1,
50
  "scale_embedding": false,
51
  "task_specific_params": {
 
69
  }
70
  },
71
  "torch_dtype": "float32",
72
+ "transformers_version": "4.40.0",
73
  "use_cache": true,
74
  "vocab_size": 50265
75
  }
generation_config.json CHANGED
@@ -4,9 +4,10 @@
4
  "decoder_start_token_id": 2,
5
  "early_stopping": true,
6
  "eos_token_id": 2,
 
7
  "forced_eos_token_id": 2,
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 1,
11
- "transformers_version": "4.39.3"
12
  }
 
4
  "decoder_start_token_id": 2,
5
  "early_stopping": true,
6
  "eos_token_id": 2,
7
+ "forced_bos_token_id": 0,
8
  "forced_eos_token_id": 2,
9
  "no_repeat_ngram_size": 3,
10
  "num_beams": 4,
11
  "pad_token_id": 1,
12
+ "transformers_version": "4.40.0"
13
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:71d3660612eac61fd3c9439d13e76fc37ccc17e1bf45f76f3aa92bb328adeee3
3
- size 1625426996
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9b006a61296ba7cd9faa3d86dc5775434f2d73b3b81c066453a01e3663cf0cd
3
+ size 557912620
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:49e927aec88c6d0faa58a11c25fc33dc62ce340cf5a82384c7f699b486b6775a
3
- size 4856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:307b444e6a871e948d060ad36b9c95aa71d7381cf25ce3ecff63bddc389e24f4
3
+ size 4920