lenatr99 commited on
Commit
b800e34
·
verified ·
1 Parent(s): 80c48d2

lora_fine_tuned_cb

Browse files
README.md CHANGED
@@ -1,9 +1,9 @@
1
  ---
2
- license: apache-2.0
3
  library_name: peft
4
  tags:
5
  - generated_from_trainer
6
- base_model: google-bert/bert-base-uncased
7
  metrics:
8
  - accuracy
9
  - f1
@@ -17,11 +17,11 @@ should probably proofread and complete it, then remove this comment. -->
17
 
18
  # lora_fine_tuned_cb
19
 
20
- This model is a fine-tuned version of [google-bert/bert-base-uncased](https://huggingface.co/google-bert/bert-base-uncased) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 1.4310
23
- - Accuracy: 0.3182
24
- - F1: 0.1536
25
 
26
  ## Model description
27
 
@@ -52,14 +52,14 @@ The following hyperparameters were used during training:
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
54
  |:-------------:|:-------:|:----:|:---------------:|:--------:|:------:|
55
- | 0.8568 | 3.5714 | 50 | 2.3106 | 0.3182 | 0.1536 |
56
- | 0.8465 | 7.1429 | 100 | 1.3703 | 0.3182 | 0.1536 |
57
- | 0.7884 | 10.7143 | 150 | 1.3701 | 0.3182 | 0.1536 |
58
- | 0.7275 | 14.2857 | 200 | 1.5024 | 0.3182 | 0.1536 |
59
- | 0.7691 | 17.8571 | 250 | 1.3990 | 0.3182 | 0.1536 |
60
- | 0.742 | 21.4286 | 300 | 1.3898 | 0.3182 | 0.1536 |
61
- | 0.7094 | 25.0 | 350 | 1.4059 | 0.3182 | 0.1536 |
62
- | 0.7238 | 28.5714 | 400 | 1.4310 | 0.3182 | 0.1536 |
63
 
64
 
65
  ### Framework versions
 
1
  ---
2
+ license: cc-by-4.0
3
  library_name: peft
4
  tags:
5
  - generated_from_trainer
6
+ base_model: EMBEDDIA/crosloengual-bert
7
  metrics:
8
  - accuracy
9
  - f1
 
17
 
18
  # lora_fine_tuned_cb
19
 
20
+ This model is a fine-tuned version of [EMBEDDIA/crosloengual-bert](https://huggingface.co/EMBEDDIA/crosloengual-bert) on an unknown dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 3.7255
23
+ - Accuracy: 0.5
24
+ - F1: 0.4690
25
 
26
  ## Model description
27
 
 
52
 
53
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 |
54
  |:-------------:|:-------:|:----:|:---------------:|:--------:|:------:|
55
+ | 0.7 | 3.5714 | 50 | 1.2589 | 0.4091 | 0.3098 |
56
+ | 0.2825 | 7.1429 | 100 | 2.2071 | 0.4091 | 0.3573 |
57
+ | 0.0229 | 10.7143 | 150 | 2.0270 | 0.6364 | 0.5916 |
58
+ | 0.0153 | 14.2857 | 200 | 2.2568 | 0.5455 | 0.5181 |
59
+ | 0.0017 | 17.8571 | 250 | 3.8454 | 0.4091 | 0.3573 |
60
+ | 0.0001 | 21.4286 | 300 | 3.8920 | 0.4091 | 0.3573 |
61
+ | 0.0 | 25.0 | 350 | 3.7418 | 0.5 | 0.4690 |
62
+ | 0.0 | 28.5714 | 400 | 3.7255 | 0.5 | 0.4690 |
63
 
64
 
65
  ### Framework versions
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "google-bert/bert-base-uncased",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -23,8 +23,8 @@
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
- "query",
27
- "value"
28
  ],
29
  "task_type": "SEQ_CLS",
30
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "EMBEDDIA/crosloengual-bert",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
23
  "rank_pattern": {},
24
  "revision": null,
25
  "target_modules": [
26
+ "value",
27
+ "query"
28
  ],
29
  "task_type": "SEQ_CLS",
30
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1deb9d7069b2ac28e0a8f637c6839c690afe32f78f9497b08381cde8b8aa4143
3
  size 2375492
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f585c81416db2e0d641ad3070fec3bb12c7441b1c21ce16c650c4a2a009e77b9
3
  size 2375492
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -8,7 +8,7 @@
8
  "single_word": false,
9
  "special": true
10
  },
11
- "100": {
12
  "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
@@ -16,7 +16,7 @@
16
  "single_word": false,
17
  "special": true
18
  },
19
- "101": {
20
  "content": "[CLS]",
21
  "lstrip": false,
22
  "normalized": false,
@@ -24,7 +24,7 @@
24
  "single_word": false,
25
  "special": true
26
  },
27
- "102": {
28
  "content": "[SEP]",
29
  "lstrip": false,
30
  "normalized": false,
@@ -32,7 +32,7 @@
32
  "single_word": false,
33
  "special": true
34
  },
35
- "103": {
36
  "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
@@ -43,9 +43,12 @@
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
- "do_lower_case": true,
 
47
  "mask_token": "[MASK]",
 
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
50
  "sep_token": "[SEP]",
51
  "strip_accents": null,
 
8
  "single_word": false,
9
  "special": true
10
  },
11
+ "102": {
12
  "content": "[UNK]",
13
  "lstrip": false,
14
  "normalized": false,
 
16
  "single_word": false,
17
  "special": true
18
  },
19
+ "103": {
20
  "content": "[CLS]",
21
  "lstrip": false,
22
  "normalized": false,
 
24
  "single_word": false,
25
  "special": true
26
  },
27
+ "104": {
28
  "content": "[SEP]",
29
  "lstrip": false,
30
  "normalized": false,
 
32
  "single_word": false,
33
  "special": true
34
  },
35
+ "105": {
36
  "content": "[MASK]",
37
  "lstrip": false,
38
  "normalized": false,
 
43
  },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
+ "do_basic_tokenize": true,
47
+ "do_lower_case": false,
48
  "mask_token": "[MASK]",
49
+ "max_len": 512,
50
  "model_max_length": 512,
51
+ "never_split": null,
52
  "pad_token": "[PAD]",
53
  "sep_token": "[SEP]",
54
  "strip_accents": null,
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1113cfb80269d7333f83921d7d077f6874ec7a5b0126440c29b610f1a6a4ebdb
3
  size 4984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c210e796accd9a980cecb92052f8aa280a5ef85100f92bbc75ceed5294509697
3
  size 4984
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff