Upload folder using huggingface_hub
Browse files- checkpoint-17632/config.json +46 -0
- checkpoint-17632/model.safetensors +3 -0
- checkpoint-17632/optimizer.pt +3 -0
- checkpoint-17632/rng_state_0.pth +3 -0
- checkpoint-17632/rng_state_1.pth +3 -0
- checkpoint-17632/rng_state_2.pth +3 -0
- checkpoint-17632/rng_state_3.pth +3 -0
- checkpoint-17632/rng_state_4.pth +3 -0
- checkpoint-17632/rng_state_5.pth +3 -0
- checkpoint-17632/rng_state_6.pth +3 -0
- checkpoint-17632/rng_state_7.pth +3 -0
- checkpoint-17632/scheduler.pt +3 -0
- checkpoint-17632/trainer_state.json +0 -0
- checkpoint-17632/training_args.bin +3 -0
- checkpoint-26448/config.json +46 -0
- checkpoint-26448/model.safetensors +3 -0
- checkpoint-26448/optimizer.pt +3 -0
- checkpoint-26448/rng_state_0.pth +3 -0
- checkpoint-26448/rng_state_1.pth +3 -0
- checkpoint-26448/rng_state_2.pth +3 -0
- checkpoint-26448/rng_state_3.pth +3 -0
- checkpoint-26448/rng_state_4.pth +3 -0
- checkpoint-26448/rng_state_5.pth +3 -0
- checkpoint-26448/rng_state_6.pth +3 -0
- checkpoint-26448/rng_state_7.pth +3 -0
- checkpoint-26448/scheduler.pt +3 -0
- checkpoint-26448/trainer_state.json +0 -0
- checkpoint-26448/training_args.bin +3 -0
- checkpoint-8816/config.json +46 -0
- checkpoint-8816/model.safetensors +3 -0
- checkpoint-8816/optimizer.pt +3 -0
- checkpoint-8816/rng_state_0.pth +3 -0
- checkpoint-8816/rng_state_1.pth +3 -0
- checkpoint-8816/rng_state_2.pth +3 -0
- checkpoint-8816/rng_state_3.pth +3 -0
- checkpoint-8816/rng_state_4.pth +3 -0
- checkpoint-8816/rng_state_5.pth +3 -0
- checkpoint-8816/rng_state_6.pth +3 -0
- checkpoint-8816/rng_state_7.pth +3 -0
- checkpoint-8816/scheduler.pt +3 -0
- checkpoint-8816/trainer_state.json +0 -0
- checkpoint-8816/training_args.bin +3 -0
- config.json +46 -0
- model.safetensors +3 -0
- training_args.bin +3 -0
checkpoint-17632/config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Qwen/Qwen2.5-0.5B",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151643,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 896,
|
11 |
+
"id2label": {
|
12 |
+
"0": "Race/Origin",
|
13 |
+
"1": "Gender/Sex",
|
14 |
+
"2": "Religion",
|
15 |
+
"3": "Ability",
|
16 |
+
"4": "Violence"
|
17 |
+
},
|
18 |
+
"initializer_range": 0.02,
|
19 |
+
"intermediate_size": 4864,
|
20 |
+
"label2id": {
|
21 |
+
"Ability": 3,
|
22 |
+
"Gender/Sex": 1,
|
23 |
+
"Race/Origin": 0,
|
24 |
+
"Religion": 2,
|
25 |
+
"Violence": 4
|
26 |
+
},
|
27 |
+
"max_position_embeddings": 32768,
|
28 |
+
"max_window_layers": 24,
|
29 |
+
"model_type": "qwen2",
|
30 |
+
"num_attention_heads": 14,
|
31 |
+
"num_hidden_layers": 24,
|
32 |
+
"num_key_value_heads": 2,
|
33 |
+
"pad_token_id": 151643,
|
34 |
+
"problem_type": "multi_label_classification",
|
35 |
+
"rms_norm_eps": 1e-06,
|
36 |
+
"rope_scaling": null,
|
37 |
+
"rope_theta": 1000000.0,
|
38 |
+
"sliding_window": null,
|
39 |
+
"tie_word_embeddings": true,
|
40 |
+
"torch_dtype": "bfloat16",
|
41 |
+
"transformers_version": "4.46.3",
|
42 |
+
"use_cache": true,
|
43 |
+
"use_mrope": false,
|
44 |
+
"use_sliding_window": false,
|
45 |
+
"vocab_size": 151936
|
46 |
+
}
|
checkpoint-17632/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a67bbe0f9fbf81d6bd961d1763e753fe482607a2f23f90330d49837ade394fd
|
3 |
+
size 988106872
|
checkpoint-17632/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:178603b76f994a604d75c89e128f005c7355dd18910cfedda4644bb505656aed
|
3 |
+
size 1976392954
|
checkpoint-17632/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06fea830cf5ad73ec00d500ea6fb952740ac936f18e93fa2d32abde1ea3ead92
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:be561d1df19be227394d8ea607c54262a06c9bf880af0aa5e04a52596a2a6cb0
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:03f3e24417a59435f5a8450a4aeb0f09cc92734b5c3b45a0701b2c043c415c05
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2bea02744c29f30024590ab1629a0e7b7dabbf1e8476456c2e7c5ce46dc35c28
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:041be966454b60c86af576fc1eb7f34189114689abff8f9622b947110f7334c8
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b85766f6596d15a810177d77dd259d9b50588cf100ec5f8ebff5fed881d57957
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8be75d04b1ebe614241b88fd010a5dda1b7bf703c00c6ebe310ca07975830fe7
|
3 |
+
size 15984
|
checkpoint-17632/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4699833a7ab4cb692996ef7567f934c0bac79d6a067963a873f89a38e412bd48
|
3 |
+
size 15984
|
checkpoint-17632/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f3fe56de841be89afc99d49fbf5620536706a75b08606c625db59355d4631eac
|
3 |
+
size 1064
|
checkpoint-17632/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-17632/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
|
3 |
+
size 5304
|
checkpoint-26448/config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Qwen/Qwen2.5-0.5B",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151643,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 896,
|
11 |
+
"id2label": {
|
12 |
+
"0": "Race/Origin",
|
13 |
+
"1": "Gender/Sex",
|
14 |
+
"2": "Religion",
|
15 |
+
"3": "Ability",
|
16 |
+
"4": "Violence"
|
17 |
+
},
|
18 |
+
"initializer_range": 0.02,
|
19 |
+
"intermediate_size": 4864,
|
20 |
+
"label2id": {
|
21 |
+
"Ability": 3,
|
22 |
+
"Gender/Sex": 1,
|
23 |
+
"Race/Origin": 0,
|
24 |
+
"Religion": 2,
|
25 |
+
"Violence": 4
|
26 |
+
},
|
27 |
+
"max_position_embeddings": 32768,
|
28 |
+
"max_window_layers": 24,
|
29 |
+
"model_type": "qwen2",
|
30 |
+
"num_attention_heads": 14,
|
31 |
+
"num_hidden_layers": 24,
|
32 |
+
"num_key_value_heads": 2,
|
33 |
+
"pad_token_id": 151643,
|
34 |
+
"problem_type": "multi_label_classification",
|
35 |
+
"rms_norm_eps": 1e-06,
|
36 |
+
"rope_scaling": null,
|
37 |
+
"rope_theta": 1000000.0,
|
38 |
+
"sliding_window": null,
|
39 |
+
"tie_word_embeddings": true,
|
40 |
+
"torch_dtype": "bfloat16",
|
41 |
+
"transformers_version": "4.46.3",
|
42 |
+
"use_cache": true,
|
43 |
+
"use_mrope": false,
|
44 |
+
"use_sliding_window": false,
|
45 |
+
"vocab_size": 151936
|
46 |
+
}
|
checkpoint-26448/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9b949843362c87d62624d1716658946c61a3295320105783f10c778f46732b21
|
3 |
+
size 988106872
|
checkpoint-26448/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bcc48a8135f22c92393584dbfcbcc54bf78b131e0ae7e03a50410a993d492f4e
|
3 |
+
size 1976392954
|
checkpoint-26448/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb18ac8d6db3307b1c242f7cb069fc8b8dab957434ddfcafcac997cfd6a43abf
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4bdab708057b5f34a402d9a2b4443f5f93a8e8ee2ddb66d955f0a15ad394ecc5
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:599882a30c163a5a2a000c4e74b320ecc4a55aa1b079882fd66aa3d2559d19e7
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:567c3b482c209c2778fc017e39a38642c488edda20673ef29f571ef7177ad81e
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0f9ffe9a916e778423aaed4ec842923c9ccfdd3d7a4fbad10dc6a3bfc278fb8e
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7ede8a81aa3c780fb9c3cb57537752a782c4aed1dcecb7aafd6ca5a7ea90252
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5b1c5c0c0afa907d332467e631e6cee80ba476689aa0caa77689ca273d83b3e4
|
3 |
+
size 15984
|
checkpoint-26448/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:73025ac422abb13303ee974109cf39f6f848de7f7013e828d04aa4e2ec0e6757
|
3 |
+
size 15984
|
checkpoint-26448/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:52793bbc3b23376c40197876c9c8d05b7549b25bd80863212845dc490717e9ca
|
3 |
+
size 1064
|
checkpoint-26448/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-26448/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
|
3 |
+
size 5304
|
checkpoint-8816/config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Qwen/Qwen2.5-0.5B",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151643,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 896,
|
11 |
+
"id2label": {
|
12 |
+
"0": "Race/Origin",
|
13 |
+
"1": "Gender/Sex",
|
14 |
+
"2": "Religion",
|
15 |
+
"3": "Ability",
|
16 |
+
"4": "Violence"
|
17 |
+
},
|
18 |
+
"initializer_range": 0.02,
|
19 |
+
"intermediate_size": 4864,
|
20 |
+
"label2id": {
|
21 |
+
"Ability": 3,
|
22 |
+
"Gender/Sex": 1,
|
23 |
+
"Race/Origin": 0,
|
24 |
+
"Religion": 2,
|
25 |
+
"Violence": 4
|
26 |
+
},
|
27 |
+
"max_position_embeddings": 32768,
|
28 |
+
"max_window_layers": 24,
|
29 |
+
"model_type": "qwen2",
|
30 |
+
"num_attention_heads": 14,
|
31 |
+
"num_hidden_layers": 24,
|
32 |
+
"num_key_value_heads": 2,
|
33 |
+
"pad_token_id": 151643,
|
34 |
+
"problem_type": "multi_label_classification",
|
35 |
+
"rms_norm_eps": 1e-06,
|
36 |
+
"rope_scaling": null,
|
37 |
+
"rope_theta": 1000000.0,
|
38 |
+
"sliding_window": null,
|
39 |
+
"tie_word_embeddings": true,
|
40 |
+
"torch_dtype": "bfloat16",
|
41 |
+
"transformers_version": "4.46.3",
|
42 |
+
"use_cache": true,
|
43 |
+
"use_mrope": false,
|
44 |
+
"use_sliding_window": false,
|
45 |
+
"vocab_size": 151936
|
46 |
+
}
|
checkpoint-8816/model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee76d4092d2ba6e5cc1af833d7e2988b9fc0cd893ecbb319b9d627ab3b716565
|
3 |
+
size 988106872
|
checkpoint-8816/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69d84b21b2cbedfa2082c54d42c9169ba44e98faed224c6270d8bceefe346831
|
3 |
+
size 1976392954
|
checkpoint-8816/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:74386f26f36ed67f56395205881e5db2d0c28ffcbeed50dd95b28771d2dac588
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:41c88f9de084200454883a13c3717941ea3fd433e2f8735507fc30611f9c5501
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:965b00d4cb4710ebab57c8787b9925bb3f77b8eeba94a186ec4bc1c2f326ef3f
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d5dc374b8b9a4c45c950f9d136feab85a767081fa59f0c7d68ed3a62060c4949
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c7c212fb779217f1edac0baf44f67b608eefc1e0e4e3f5a9dd7eb557032c1bc
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86e1effd626ce1e95dd68a0c8089fe19218f2b24dfe9e45ed2cab1c0ebc10ba1
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:799cc83f60dfc1c4243cfd6403592112414a2eec494e6832f10221c96ff62c20
|
3 |
+
size 15984
|
checkpoint-8816/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:586777c398770c3255d3a1f48c7fef44ea9d89117c627c9ea490e16bfd9a49ba
|
3 |
+
size 15984
|
checkpoint-8816/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ca8a5e3f861b24accec8b677236a58d72e7cfbd9bc6c73440608bc95a509cee2
|
3 |
+
size 1064
|
checkpoint-8816/trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
checkpoint-8816/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
|
3 |
+
size 5304
|
config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "Qwen/Qwen2.5-0.5B",
|
3 |
+
"architectures": [
|
4 |
+
"Qwen2ForSequenceClassification"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bos_token_id": 151643,
|
8 |
+
"eos_token_id": 151643,
|
9 |
+
"hidden_act": "silu",
|
10 |
+
"hidden_size": 896,
|
11 |
+
"id2label": {
|
12 |
+
"0": "Race/Origin",
|
13 |
+
"1": "Gender/Sex",
|
14 |
+
"2": "Religion",
|
15 |
+
"3": "Ability",
|
16 |
+
"4": "Violence"
|
17 |
+
},
|
18 |
+
"initializer_range": 0.02,
|
19 |
+
"intermediate_size": 4864,
|
20 |
+
"label2id": {
|
21 |
+
"Ability": 3,
|
22 |
+
"Gender/Sex": 1,
|
23 |
+
"Race/Origin": 0,
|
24 |
+
"Religion": 2,
|
25 |
+
"Violence": 4
|
26 |
+
},
|
27 |
+
"max_position_embeddings": 32768,
|
28 |
+
"max_window_layers": 24,
|
29 |
+
"model_type": "qwen2",
|
30 |
+
"num_attention_heads": 14,
|
31 |
+
"num_hidden_layers": 24,
|
32 |
+
"num_key_value_heads": 2,
|
33 |
+
"pad_token_id": 151643,
|
34 |
+
"problem_type": "multi_label_classification",
|
35 |
+
"rms_norm_eps": 1e-06,
|
36 |
+
"rope_scaling": null,
|
37 |
+
"rope_theta": 1000000.0,
|
38 |
+
"sliding_window": null,
|
39 |
+
"tie_word_embeddings": true,
|
40 |
+
"torch_dtype": "bfloat16",
|
41 |
+
"transformers_version": "4.46.3",
|
42 |
+
"use_cache": true,
|
43 |
+
"use_mrope": false,
|
44 |
+
"use_sliding_window": false,
|
45 |
+
"vocab_size": 151936
|
46 |
+
}
|
model.safetensors
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ee76d4092d2ba6e5cc1af833d7e2988b9fc0cd893ecbb319b9d627ab3b716565
|
3 |
+
size 988106872
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
|
3 |
+
size 5304
|