tcapelle commited on
Commit
c1d0c89
·
verified ·
1 Parent(s): 36b6c53

Upload folder using huggingface_hub

Browse files
Files changed (45) hide show
  1. checkpoint-17632/config.json +46 -0
  2. checkpoint-17632/model.safetensors +3 -0
  3. checkpoint-17632/optimizer.pt +3 -0
  4. checkpoint-17632/rng_state_0.pth +3 -0
  5. checkpoint-17632/rng_state_1.pth +3 -0
  6. checkpoint-17632/rng_state_2.pth +3 -0
  7. checkpoint-17632/rng_state_3.pth +3 -0
  8. checkpoint-17632/rng_state_4.pth +3 -0
  9. checkpoint-17632/rng_state_5.pth +3 -0
  10. checkpoint-17632/rng_state_6.pth +3 -0
  11. checkpoint-17632/rng_state_7.pth +3 -0
  12. checkpoint-17632/scheduler.pt +3 -0
  13. checkpoint-17632/trainer_state.json +0 -0
  14. checkpoint-17632/training_args.bin +3 -0
  15. checkpoint-26448/config.json +46 -0
  16. checkpoint-26448/model.safetensors +3 -0
  17. checkpoint-26448/optimizer.pt +3 -0
  18. checkpoint-26448/rng_state_0.pth +3 -0
  19. checkpoint-26448/rng_state_1.pth +3 -0
  20. checkpoint-26448/rng_state_2.pth +3 -0
  21. checkpoint-26448/rng_state_3.pth +3 -0
  22. checkpoint-26448/rng_state_4.pth +3 -0
  23. checkpoint-26448/rng_state_5.pth +3 -0
  24. checkpoint-26448/rng_state_6.pth +3 -0
  25. checkpoint-26448/rng_state_7.pth +3 -0
  26. checkpoint-26448/scheduler.pt +3 -0
  27. checkpoint-26448/trainer_state.json +0 -0
  28. checkpoint-26448/training_args.bin +3 -0
  29. checkpoint-8816/config.json +46 -0
  30. checkpoint-8816/model.safetensors +3 -0
  31. checkpoint-8816/optimizer.pt +3 -0
  32. checkpoint-8816/rng_state_0.pth +3 -0
  33. checkpoint-8816/rng_state_1.pth +3 -0
  34. checkpoint-8816/rng_state_2.pth +3 -0
  35. checkpoint-8816/rng_state_3.pth +3 -0
  36. checkpoint-8816/rng_state_4.pth +3 -0
  37. checkpoint-8816/rng_state_5.pth +3 -0
  38. checkpoint-8816/rng_state_6.pth +3 -0
  39. checkpoint-8816/rng_state_7.pth +3 -0
  40. checkpoint-8816/scheduler.pt +3 -0
  41. checkpoint-8816/trainer_state.json +0 -0
  42. checkpoint-8816/training_args.bin +3 -0
  43. config.json +46 -0
  44. model.safetensors +3 -0
  45. training_args.bin +3 -0
checkpoint-17632/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2.5-0.5B",
3
+ "architectures": [
4
+ "Qwen2ForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 896,
11
+ "id2label": {
12
+ "0": "Race/Origin",
13
+ "1": "Gender/Sex",
14
+ "2": "Religion",
15
+ "3": "Ability",
16
+ "4": "Violence"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 4864,
20
+ "label2id": {
21
+ "Ability": 3,
22
+ "Gender/Sex": 1,
23
+ "Race/Origin": 0,
24
+ "Religion": 2,
25
+ "Violence": 4
26
+ },
27
+ "max_position_embeddings": 32768,
28
+ "max_window_layers": 24,
29
+ "model_type": "qwen2",
30
+ "num_attention_heads": 14,
31
+ "num_hidden_layers": 24,
32
+ "num_key_value_heads": 2,
33
+ "pad_token_id": 151643,
34
+ "problem_type": "multi_label_classification",
35
+ "rms_norm_eps": 1e-06,
36
+ "rope_scaling": null,
37
+ "rope_theta": 1000000.0,
38
+ "sliding_window": null,
39
+ "tie_word_embeddings": true,
40
+ "torch_dtype": "bfloat16",
41
+ "transformers_version": "4.46.3",
42
+ "use_cache": true,
43
+ "use_mrope": false,
44
+ "use_sliding_window": false,
45
+ "vocab_size": 151936
46
+ }
checkpoint-17632/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a67bbe0f9fbf81d6bd961d1763e753fe482607a2f23f90330d49837ade394fd
3
+ size 988106872
checkpoint-17632/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:178603b76f994a604d75c89e128f005c7355dd18910cfedda4644bb505656aed
3
+ size 1976392954
checkpoint-17632/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06fea830cf5ad73ec00d500ea6fb952740ac936f18e93fa2d32abde1ea3ead92
3
+ size 15984
checkpoint-17632/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be561d1df19be227394d8ea607c54262a06c9bf880af0aa5e04a52596a2a6cb0
3
+ size 15984
checkpoint-17632/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:03f3e24417a59435f5a8450a4aeb0f09cc92734b5c3b45a0701b2c043c415c05
3
+ size 15984
checkpoint-17632/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2bea02744c29f30024590ab1629a0e7b7dabbf1e8476456c2e7c5ce46dc35c28
3
+ size 15984
checkpoint-17632/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:041be966454b60c86af576fc1eb7f34189114689abff8f9622b947110f7334c8
3
+ size 15984
checkpoint-17632/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b85766f6596d15a810177d77dd259d9b50588cf100ec5f8ebff5fed881d57957
3
+ size 15984
checkpoint-17632/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8be75d04b1ebe614241b88fd010a5dda1b7bf703c00c6ebe310ca07975830fe7
3
+ size 15984
checkpoint-17632/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4699833a7ab4cb692996ef7567f934c0bac79d6a067963a873f89a38e412bd48
3
+ size 15984
checkpoint-17632/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3fe56de841be89afc99d49fbf5620536706a75b08606c625db59355d4631eac
3
+ size 1064
checkpoint-17632/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-17632/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
3
+ size 5304
checkpoint-26448/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2.5-0.5B",
3
+ "architectures": [
4
+ "Qwen2ForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 896,
11
+ "id2label": {
12
+ "0": "Race/Origin",
13
+ "1": "Gender/Sex",
14
+ "2": "Religion",
15
+ "3": "Ability",
16
+ "4": "Violence"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 4864,
20
+ "label2id": {
21
+ "Ability": 3,
22
+ "Gender/Sex": 1,
23
+ "Race/Origin": 0,
24
+ "Religion": 2,
25
+ "Violence": 4
26
+ },
27
+ "max_position_embeddings": 32768,
28
+ "max_window_layers": 24,
29
+ "model_type": "qwen2",
30
+ "num_attention_heads": 14,
31
+ "num_hidden_layers": 24,
32
+ "num_key_value_heads": 2,
33
+ "pad_token_id": 151643,
34
+ "problem_type": "multi_label_classification",
35
+ "rms_norm_eps": 1e-06,
36
+ "rope_scaling": null,
37
+ "rope_theta": 1000000.0,
38
+ "sliding_window": null,
39
+ "tie_word_embeddings": true,
40
+ "torch_dtype": "bfloat16",
41
+ "transformers_version": "4.46.3",
42
+ "use_cache": true,
43
+ "use_mrope": false,
44
+ "use_sliding_window": false,
45
+ "vocab_size": 151936
46
+ }
checkpoint-26448/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b949843362c87d62624d1716658946c61a3295320105783f10c778f46732b21
3
+ size 988106872
checkpoint-26448/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bcc48a8135f22c92393584dbfcbcc54bf78b131e0ae7e03a50410a993d492f4e
3
+ size 1976392954
checkpoint-26448/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb18ac8d6db3307b1c242f7cb069fc8b8dab957434ddfcafcac997cfd6a43abf
3
+ size 15984
checkpoint-26448/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4bdab708057b5f34a402d9a2b4443f5f93a8e8ee2ddb66d955f0a15ad394ecc5
3
+ size 15984
checkpoint-26448/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:599882a30c163a5a2a000c4e74b320ecc4a55aa1b079882fd66aa3d2559d19e7
3
+ size 15984
checkpoint-26448/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:567c3b482c209c2778fc017e39a38642c488edda20673ef29f571ef7177ad81e
3
+ size 15984
checkpoint-26448/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f9ffe9a916e778423aaed4ec842923c9ccfdd3d7a4fbad10dc6a3bfc278fb8e
3
+ size 15984
checkpoint-26448/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7ede8a81aa3c780fb9c3cb57537752a782c4aed1dcecb7aafd6ca5a7ea90252
3
+ size 15984
checkpoint-26448/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b1c5c0c0afa907d332467e631e6cee80ba476689aa0caa77689ca273d83b3e4
3
+ size 15984
checkpoint-26448/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73025ac422abb13303ee974109cf39f6f848de7f7013e828d04aa4e2ec0e6757
3
+ size 15984
checkpoint-26448/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52793bbc3b23376c40197876c9c8d05b7549b25bd80863212845dc490717e9ca
3
+ size 1064
checkpoint-26448/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-26448/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
3
+ size 5304
checkpoint-8816/config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2.5-0.5B",
3
+ "architectures": [
4
+ "Qwen2ForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 896,
11
+ "id2label": {
12
+ "0": "Race/Origin",
13
+ "1": "Gender/Sex",
14
+ "2": "Religion",
15
+ "3": "Ability",
16
+ "4": "Violence"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 4864,
20
+ "label2id": {
21
+ "Ability": 3,
22
+ "Gender/Sex": 1,
23
+ "Race/Origin": 0,
24
+ "Religion": 2,
25
+ "Violence": 4
26
+ },
27
+ "max_position_embeddings": 32768,
28
+ "max_window_layers": 24,
29
+ "model_type": "qwen2",
30
+ "num_attention_heads": 14,
31
+ "num_hidden_layers": 24,
32
+ "num_key_value_heads": 2,
33
+ "pad_token_id": 151643,
34
+ "problem_type": "multi_label_classification",
35
+ "rms_norm_eps": 1e-06,
36
+ "rope_scaling": null,
37
+ "rope_theta": 1000000.0,
38
+ "sliding_window": null,
39
+ "tie_word_embeddings": true,
40
+ "torch_dtype": "bfloat16",
41
+ "transformers_version": "4.46.3",
42
+ "use_cache": true,
43
+ "use_mrope": false,
44
+ "use_sliding_window": false,
45
+ "vocab_size": 151936
46
+ }
checkpoint-8816/model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee76d4092d2ba6e5cc1af833d7e2988b9fc0cd893ecbb319b9d627ab3b716565
3
+ size 988106872
checkpoint-8816/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69d84b21b2cbedfa2082c54d42c9169ba44e98faed224c6270d8bceefe346831
3
+ size 1976392954
checkpoint-8816/rng_state_0.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:74386f26f36ed67f56395205881e5db2d0c28ffcbeed50dd95b28771d2dac588
3
+ size 15984
checkpoint-8816/rng_state_1.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:41c88f9de084200454883a13c3717941ea3fd433e2f8735507fc30611f9c5501
3
+ size 15984
checkpoint-8816/rng_state_2.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:965b00d4cb4710ebab57c8787b9925bb3f77b8eeba94a186ec4bc1c2f326ef3f
3
+ size 15984
checkpoint-8816/rng_state_3.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d5dc374b8b9a4c45c950f9d136feab85a767081fa59f0c7d68ed3a62060c4949
3
+ size 15984
checkpoint-8816/rng_state_4.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c7c212fb779217f1edac0baf44f67b608eefc1e0e4e3f5a9dd7eb557032c1bc
3
+ size 15984
checkpoint-8816/rng_state_5.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86e1effd626ce1e95dd68a0c8089fe19218f2b24dfe9e45ed2cab1c0ebc10ba1
3
+ size 15984
checkpoint-8816/rng_state_6.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:799cc83f60dfc1c4243cfd6403592112414a2eec494e6832f10221c96ff62c20
3
+ size 15984
checkpoint-8816/rng_state_7.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:586777c398770c3255d3a1f48c7fef44ea9d89117c627c9ea490e16bfd9a49ba
3
+ size 15984
checkpoint-8816/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ca8a5e3f861b24accec8b677236a58d72e7cfbd9bc6c73440608bc95a509cee2
3
+ size 1064
checkpoint-8816/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
checkpoint-8816/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
3
+ size 5304
config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "Qwen/Qwen2.5-0.5B",
3
+ "architectures": [
4
+ "Qwen2ForSequenceClassification"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 896,
11
+ "id2label": {
12
+ "0": "Race/Origin",
13
+ "1": "Gender/Sex",
14
+ "2": "Religion",
15
+ "3": "Ability",
16
+ "4": "Violence"
17
+ },
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 4864,
20
+ "label2id": {
21
+ "Ability": 3,
22
+ "Gender/Sex": 1,
23
+ "Race/Origin": 0,
24
+ "Religion": 2,
25
+ "Violence": 4
26
+ },
27
+ "max_position_embeddings": 32768,
28
+ "max_window_layers": 24,
29
+ "model_type": "qwen2",
30
+ "num_attention_heads": 14,
31
+ "num_hidden_layers": 24,
32
+ "num_key_value_heads": 2,
33
+ "pad_token_id": 151643,
34
+ "problem_type": "multi_label_classification",
35
+ "rms_norm_eps": 1e-06,
36
+ "rope_scaling": null,
37
+ "rope_theta": 1000000.0,
38
+ "sliding_window": null,
39
+ "tie_word_embeddings": true,
40
+ "torch_dtype": "bfloat16",
41
+ "transformers_version": "4.46.3",
42
+ "use_cache": true,
43
+ "use_mrope": false,
44
+ "use_sliding_window": false,
45
+ "vocab_size": 151936
46
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee76d4092d2ba6e5cc1af833d7e2988b9fc0cd893ecbb319b9d627ab3b716565
3
+ size 988106872
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:71c02dfc99ce114336ae3e1782c3d8eca3ebdc4417ad299a59b676617d90d71c
3
+ size 5304