charisgao commited on
Commit
2cb840f
·
verified ·
1 Parent(s): a519894

5300-biased-word-detection

Browse files
Files changed (5) hide show
  1. README.md +11 -11
  2. config.json +6 -3
  3. model.safetensors +2 -2
  4. special_tokens_map.json +6 -42
  5. tokenizer.json +6 -1
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: transformers
3
  license: mit
4
- base_model: charisgao/pretrain_model
5
  tags:
6
  - generated_from_trainer
7
  metrics:
@@ -19,13 +19,13 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  # model
21
 
22
- This model is a fine-tuned version of [charisgao/pretrain_model](https://huggingface.co/charisgao/pretrain_model) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
- - Loss: 0.5988
25
- - Precision: 0.7906
26
- - Recall: 0.925
27
- - F1: 0.8525
28
- - Accuracy: 0.7942
29
 
30
  ## Model description
31
 
@@ -56,14 +56,14 @@ The following hyperparameters were used during training:
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
58
  |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
59
- | 0.5484 | 0.8547 | 100 | 0.5398 | 0.7911 | 0.8725 | 0.8298 | 0.7645 |
60
- | 0.4185 | 1.7094 | 200 | 0.5482 | 0.8034 | 0.9216 | 0.8584 | 0.8 |
61
- | 0.3595 | 2.5641 | 300 | 0.6521 | 0.776 | 0.9510 | 0.8546 | 0.7871 |
62
 
63
 
64
  ### Framework versions
65
 
66
  - Transformers 4.46.3
67
  - Pytorch 2.5.1+cu121
68
- - Datasets 3.1.0
69
  - Tokenizers 0.20.3
 
1
  ---
2
  library_name: transformers
3
  license: mit
4
+ base_model: roberta-base
5
  tags:
6
  - generated_from_trainer
7
  metrics:
 
19
 
20
  # model
21
 
22
+ This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset.
23
  It achieves the following results on the evaluation set:
24
+ - Loss: 0.3309
25
+ - Precision: 0.6214
26
+ - Recall: 0.5332
27
+ - F1: 0.5410
28
+ - Accuracy: 0.9060
29
 
30
  ## Model description
31
 
 
56
 
57
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
58
  |:-------------:|:------:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
59
+ | 0.1159 | 0.8547 | 100 | 0.3248 | 0.6055 | 0.5785 | 0.5888 | 0.8852 |
60
+ | 0.202 | 1.7094 | 200 | 0.3075 | 0.6661 | 0.5444 | 0.5581 | 0.9087 |
61
+ | 0.1593 | 2.5641 | 300 | 0.3221 | 0.6624 | 0.5473 | 0.5622 | 0.9079 |
62
 
63
 
64
  ### Framework versions
65
 
66
  - Transformers 4.46.3
67
  - Pytorch 2.5.1+cu121
68
+ - Datasets 3.2.0
69
  - Tokenizers 0.20.3
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "charisgao/pretrain_model",
3
  "architectures": [
4
- "RobertaForSequenceClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
@@ -10,6 +10,10 @@
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
 
 
 
 
13
  "initializer_range": 0.02,
14
  "intermediate_size": 3072,
15
  "layer_norm_eps": 1e-05,
@@ -19,7 +23,6 @@
19
  "num_hidden_layers": 12,
20
  "pad_token_id": 1,
21
  "position_embedding_type": "absolute",
22
- "problem_type": "single_label_classification",
23
  "torch_dtype": "float32",
24
  "transformers_version": "4.46.3",
25
  "type_vocab_size": 1,
 
1
  {
2
+ "_name_or_path": "roberta-base",
3
  "architectures": [
4
+ "RobertaForTokenClassification"
5
  ],
6
  "attention_probs_dropout_prob": 0.1,
7
  "bos_token_id": 0,
 
10
  "hidden_act": "gelu",
11
  "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
13
+ "id2label": {
14
+ "0": "Non-biased",
15
+ "1": "Biased"
16
+ },
17
  "initializer_range": 0.02,
18
  "intermediate_size": 3072,
19
  "layer_norm_eps": 1e-05,
 
23
  "num_hidden_layers": 12,
24
  "pad_token_id": 1,
25
  "position_embedding_type": "absolute",
 
26
  "torch_dtype": "float32",
27
  "transformers_version": "4.46.3",
28
  "type_vocab_size": 1,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:95d3df44056589767f2483300c4a04574c4887bb2ace9f0bf32ff09deec64623
3
- size 498612824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74973004aa8b8820abde4063cc091b52506d8d9c6fe0d5f8819c330a960fbe44
3
+ size 496250232
special_tokens_map.json CHANGED
@@ -1,25 +1,7 @@
1
  {
2
- "bos_token": {
3
- "content": "<s>",
4
- "lstrip": false,
5
- "normalized": true,
6
- "rstrip": false,
7
- "single_word": false
8
- },
9
- "cls_token": {
10
- "content": "<s>",
11
- "lstrip": false,
12
- "normalized": true,
13
- "rstrip": false,
14
- "single_word": false
15
- },
16
- "eos_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": true,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "mask_token": {
24
  "content": "<mask>",
25
  "lstrip": true,
@@ -27,25 +9,7 @@
27
  "rstrip": false,
28
  "single_word": false
29
  },
30
- "pad_token": {
31
- "content": "<pad>",
32
- "lstrip": false,
33
- "normalized": true,
34
- "rstrip": false,
35
- "single_word": false
36
- },
37
- "sep_token": {
38
- "content": "</s>",
39
- "lstrip": false,
40
- "normalized": true,
41
- "rstrip": false,
42
- "single_word": false
43
- },
44
- "unk_token": {
45
- "content": "<unk>",
46
- "lstrip": false,
47
- "normalized": true,
48
- "rstrip": false,
49
- "single_word": false
50
- }
51
  }
 
1
  {
2
+ "bos_token": "<s>",
3
+ "cls_token": "<s>",
4
+ "eos_token": "</s>",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "mask_token": {
6
  "content": "<mask>",
7
  "lstrip": true,
 
9
  "rstrip": false,
10
  "single_word": false
11
  },
12
+ "pad_token": "<pad>",
13
+ "sep_token": "</s>",
14
+ "unk_token": "<unk>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
tokenizer.json CHANGED
@@ -1,6 +1,11 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
 
 
 
 
 
4
  "padding": null,
5
  "added_tokens": [
6
  {
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
  "padding": null,
10
  "added_tokens": [
11
  {