versae commited on
Commit
90c3afe
·
verified ·
1 Parent(s): 49e465b

Upload folder using huggingface_hub (#1)

Browse files

- 30b5e5c13b045c46fbd60ad2106390632816ffb5ec145f67d234bac6f78e92b8 (9f5d389aab172ae23e1c49c025c3c07ead629e93)
- a094fcd31cd8f0e391a462702a6c575d619a9614c9b6778ffe33f6de21e143b9 (284991c7768e1ce722f7f29dac749a9e08bc2148)
- b618744c54f58b56d1ea85bcb9019120703839d490305f53aaf9dd797d113c11 (72a24a38df4aca0537c042f1093fd2e95030e2a9)
- 8d28d75b3c2600e6638941e9c119e0c4abc6aa29e2e4681bf05bd211d9daee95 (8280a57c905518333279903a69b6fd6e7c18cd7d)
- fcd171f06e6d2203782388fc885cd25557f8c1db27a4ca8be41478608b4a907f (656431f44cf04c504b58414c753a54164c97f67d)
- 07b8512ee6ef61b6983c60e49e586b806e64a187b98861ec2ca43142aa3c68f2 (01a02b3a212fd3c9455c799a8ed80f49db862b52)
- 3f48285d9fca47099e42bcd1de2173ac93dce25471e02b7e00c6076f0bf34890 (c95547579bebfefa82d7c27e258fca15bf1b5dcd)
- 996b0ed3e48a4610cdc9a400a9e2bc106639cb2e4f17ab4812e4601e27487234 (ba2cf9783389bf24a8a950c132473c8c0222e324)
- 1c3af2f2dfb5d27d1d79497c701336686c833044868343c3d7f0923c68c1a56b (de23430bb7eb01203467e4c79388fb6f22652a17)
- 925e141e363719ebbb1947d4dd1bbee00e7a1fbe46b85084a3cdb949822af88a (c239f70176e99d26ca25292e0eac9a8bdb0f39a7)
- 3c1afe61e6bd1662319216ae839e7c9295c54f20a8a3386f66b2e0a4d898f94a (33cd1a607edf2b9dad00dcec08917ef36c26cd9d)
- 17934129ba8e6c8dbcb73f5a7974caf8770d6696155a810b37b1750f43c4c3ca (d12f0cb5e6b981e6e571289e11372665f7e5bb05)
- 6cd7acb9b861adc6ed49c4c8b4b163f94b426f2491974ec28f9e3ae37ca4abed (6486112fb93a2e8e15a028d712b02df62ab599b4)
- a1e9d8bc16646653b55ec512d3f16a85f8ea5f5825aafd28898c81b3d1485a75 (aae0b6d4a75ee3f1354573ec22b737f2cace3645)
- f6c87821d728c164d2d312b9ff1e93aa133e1b267d102923bfcc2b2cf0d239fe (7d9eb8e49d439ab59c4cfd6952c404dd56d9e949)
- d37749578b274ff2fdb1d54e90a6253838fa183b465fe9e39bc4f75176a64657 (56a05aac91afbe012e9fe3ee16d4b9688ad919d8)
- 490a10429d135e772a8f5def6b605571dcedafa353599cd9cbcb68e4c088516e (f782f224920b4e1b92a1441709d0174f40386fa6)
- ace13f4e7e79270ae55079dea597a9877340d3829ff0015b449088de67670790 (b28ea0b7a4cc28f75368e736e92659d4e13a32c8)
- 3f3732d2a19fb798e466dc2ac94c11c7f3144ff10e541dc4297a674de9e2428a (5208487fd198661c10e1e784ed3d62a871a6faf2)
- b006dc5a95ecb397987419c04f6359192b583f232f2b4b93d99804a8b6e5f5e3 (9f7c0fcc8638be8a973b0edf8b633aa1278cb967)
- 6a3f2a88652c5bf76b13d539f91c68c3b5043e141c8d07eef193b3ec0e2e17f8 (34ac80c00d660eb67ab6b24cc0b19e2529e547a5)
- d3540a14becff2a8f2d17b819b95a429d006dbe1b7c5c78592484a587e969f19 (95df142ba9934ec2eeb1f2f064343abbd2a40e1c)
- 56ac7c1569c613e2cf8b657c4e1725f61e5d6d66810f66c5c90680bc601b887b (9993d756e9f9446a0409ea520917dc4f1d660b33)
- 0c0e95d80f99a1e9731000027937cd107d20661a87c74526069b63e5c6b94970 (b1ffad13921009285969910aaa9a181f1b9829de)
- a4fbd6c2891a266e49af0d5ba5b33077886b5ac09d21634e9d0a73ba9e379114 (c3e6779c777b52bcc6e782d5c018ea4307e78e56)
- 846d3e4cab30d4e00f6e111a18fd1f7d01c7f2c44e399441c60b96d2518fb40c (b252ab01c1f3033d9d72dc7d0ad5ad72a59aaa56)
- 26db87ae8a66e6888cf82d9c2905675a82628a98a85480f02fdc1dd30b45c364 (0a69a14457f9c62e2ba5b9b92af2be7705307c3a)
- 1b00bad3653633e508f69f0ffc5fc66327a25261b6fc951db475ae48d71e26d6 (4087922f7327f0cdf11d2c951e11a41eee5ffe89)
- ea76e0c8cd47b339d7c938dc3f5a7f8ff62936f01e1573d393964c4484cda8ca (9053fe4feea8b4c61ff3a213eadc0f3b0ff719c1)

README.md ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - mlx
5
+ ---
6
+
7
+ # mlx-community/Mixtral-8x22B-Instruct-v0.1-8bit
8
+ This model was converted to MLX format from [`mistralai/Mixtral-8x22B-Instruct-v0.1`]() using mlx-lm version **0.9.0**.
9
+ Refer to the [original model card](https://huggingface.co/mistralai/Mixtral-8x22B-Instruct-v0.1) for more details on the model.
10
+ ## Use with mlx
11
+
12
+ ```bash
13
+ pip install mlx-lm
14
+ ```
15
+
16
+ ```python
17
+ from mlx_lm import load, generate
18
+
19
+ model, tokenizer = load("mlx-community/Mixtral-8x22B-Instruct-v0.1-8bit")
20
+ response = generate(model, tokenizer, prompt="hello", verbose=True)
21
+ ```
config.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_cross_attention": false,
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bad_words_ids": null,
8
+ "begin_suppress_tokens": null,
9
+ "bos_token_id": 1,
10
+ "chunk_size_feed_forward": 0,
11
+ "cross_attention_hidden_size": null,
12
+ "decoder_start_token_id": null,
13
+ "diversity_penalty": 0.0,
14
+ "do_sample": false,
15
+ "early_stopping": false,
16
+ "encoder_no_repeat_ngram_size": 0,
17
+ "eos_token_id": 2,
18
+ "exponential_decay_length_penalty": null,
19
+ "finetuning_task": null,
20
+ "forced_bos_token_id": null,
21
+ "forced_eos_token_id": null,
22
+ "hidden_act": "silu",
23
+ "hidden_size": 6144,
24
+ "id2label": {
25
+ "0": "LABEL_0",
26
+ "1": "LABEL_1"
27
+ },
28
+ "initializer_range": 0.02,
29
+ "intermediate_size": 16384,
30
+ "is_decoder": false,
31
+ "is_encoder_decoder": false,
32
+ "label2id": {
33
+ "LABEL_0": 0,
34
+ "LABEL_1": 1
35
+ },
36
+ "length_penalty": 1.0,
37
+ "max_length": 20,
38
+ "max_position_embeddings": 65536,
39
+ "min_length": 0,
40
+ "model_type": "mixtral",
41
+ "no_repeat_ngram_size": 0,
42
+ "num_attention_heads": 48,
43
+ "num_beam_groups": 1,
44
+ "num_beams": 1,
45
+ "num_experts_per_tok": 2,
46
+ "num_hidden_layers": 56,
47
+ "num_key_value_heads": 8,
48
+ "num_local_experts": 8,
49
+ "num_return_sequences": 1,
50
+ "output_attentions": false,
51
+ "output_hidden_states": false,
52
+ "output_router_logits": false,
53
+ "output_scores": false,
54
+ "pad_token_id": null,
55
+ "prefix": null,
56
+ "problem_type": null,
57
+ "pruned_heads": {},
58
+ "quantization": {
59
+ "group_size": 64,
60
+ "bits": 8
61
+ },
62
+ "remove_invalid_values": false,
63
+ "repetition_penalty": 1.0,
64
+ "return_dict": true,
65
+ "return_dict_in_generate": false,
66
+ "rms_norm_eps": 1e-05,
67
+ "rope_theta": 1000000.0,
68
+ "router_aux_loss_coef": 0.001,
69
+ "sep_token_id": null,
70
+ "sliding_window": null,
71
+ "suppress_tokens": null,
72
+ "task_specific_params": null,
73
+ "temperature": 1.0,
74
+ "tf_legacy_loss": false,
75
+ "tie_encoder_decoder": false,
76
+ "tie_word_embeddings": false,
77
+ "tokenizer_class": null,
78
+ "top_k": 50,
79
+ "top_p": 1.0,
80
+ "torch_dtype": "bfloat16",
81
+ "torchscript": false,
82
+ "transformers_version": "4.39.3",
83
+ "typical_p": 1.0,
84
+ "use_bfloat16": false,
85
+ "use_cache": true,
86
+ "vocab_size": 32768
87
+ }
model-00001-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb6744d54e379cc56c209b3562d78a06f6dfa187016c7beb2aaa9d21af4b676f
3
+ size 5296074524
model-00002-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:02a60d15535b1e0b429a3a8184943ea4d31233e3e6cd6d3686bdb7e776036089
3
+ size 5321266592
model-00003-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:183004cac899d86bdf0562611195604ea6a978191c43489705f293bd6096eb32
3
+ size 5321266588
model-00004-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e7aed71beed50e93ae746cbdfc92928018520cc2b9f900604383152c75b5723c
3
+ size 5321266602
model-00005-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1ad402be4b0c1fc43d39d762389b5a27bc975a3abf066c4509c390d6afd148e
3
+ size 5321266610
model-00006-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ecdccfd56a328f20704f66aa5e77f0d3b19fcae373763d4637f54a768d86b27
3
+ size 5321266732
model-00007-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9d59cc40099b308b918befd3d187d0ce0d2dcb5fed65a2c7e68efba284de9b34
3
+ size 5321266768
model-00008-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf723ad3e1a454155ee8b2214207a6028781c9accf3db4330d7c5dbfc295997b
3
+ size 5321266778
model-00009-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6928dabc1f5dc8a5778a43c3ec3c15af060514b4cf14204678a47ffaffeb8378
3
+ size 5321266776
model-00010-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:14b71df96bf6998e192143bfea21ca102a7b4a89ba979e09087856ac63018e1f
3
+ size 5321266766
model-00011-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d810396409ab21a36b4c41b63e2c70ebd6cea86db3389ba048a6faf69e4fbbf8
3
+ size 5321266770
model-00012-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fed4d33e8fb809102851e92de6e761cc9b5eefb29c3718a2f59231d3c4d4a04c
3
+ size 5321266754
model-00013-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76f06b9b4802ea1af03b44c87cf8e87371a60a3045ebc66f3ff212c41062b0a8
3
+ size 5321266720
model-00014-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:273496bc6c71c9005970ad68354be366123a4e8a85776239743238cb99cb418f
3
+ size 5321266784
model-00015-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b74fb40ae399a6171237887aefc40ee93bd5dbfad4d54d756809249cd285c5c3
3
+ size 5321266782
model-00016-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b433338d423e6e0dffb226d62e70739fb0e2f1bdcf6f76546eb5ee3288e5c03e
3
+ size 5321266766
model-00017-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ec1635af29bd9957455cf590f393a0d4db3703e3a3cb127811c6f3d4fc8e81f
3
+ size 5321266764
model-00018-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e81c6e40ea718be515c647480e34f5daec9634b8197b8bd9f2debfa09f54e5c
3
+ size 5321266778
model-00019-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d89d44e15e66e6fce9bba7240015397263ffdcecd0da8a46ee173db5db11f130
3
+ size 5321266772
model-00020-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9213616e2342b9a47085640b46e0aefee32d7a9e9e5eea840ef7815ecfde73af
3
+ size 5321266778
model-00021-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c269b5bafecd7f256fe33ba919cc7de99ab33a9d1fd6bdd666d37adb661e3e5b
3
+ size 5321266750
model-00022-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28902e02f9ac9b5655d068958c6ac7eecc7804d5c5b79fe98887da85b3944b9a
3
+ size 5321266764
model-00023-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fdcf31e0d42326a6868a1737c2063f5238d5a5f6a56a32de24fb12d16a2c4c9
3
+ size 5321266772
model-00024-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f3ebfc406d7351de656ba584fe5549a1d6cbd8eb7b7ea8f97f33224a4c858cf
3
+ size 5321266754
model-00025-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d0722fcb68d479f88df43092ac2a80b47cdc7cd00448b4645909ae3dc9c594b
3
+ size 5321266772
model-00026-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3192abd43305e47713862e37136b7e7048884a56f5a43f4ab7b116174eaa46ed
3
+ size 5321266776
model-00027-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b857e5ad08617d906024a8763daf2dc8d0ae1d47f5975a9b294b844607476f6e
3
+ size 5321266774
model-00028-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c91d80c1c9d44118eb7f15fd579e9c987b7582deb18f24a5a3b8d2b790732ab2
3
+ size 5321266768
model-00029-of-00029.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3727875b033b34db156ac0b1aacfce305343e6007c8f6fd82376e9e90e08722a
3
+ size 641767520
model.safetensors.index.json ADDED
The diff for this file is too large to render. See raw diff
 
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,99 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": true,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": true,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": true,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "3": {
30
+ "content": "[INST]",
31
+ "lstrip": false,
32
+ "normalized": true,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "4": {
38
+ "content": "[/INST]",
39
+ "lstrip": false,
40
+ "normalized": true,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "5": {
46
+ "content": "[TOOL_CALLS]",
47
+ "lstrip": false,
48
+ "normalized": true,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "6": {
54
+ "content": "[AVAILABLE_TOOLS]",
55
+ "lstrip": false,
56
+ "normalized": true,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "7": {
62
+ "content": "[/AVAILABLE_TOOLS]",
63
+ "lstrip": false,
64
+ "normalized": true,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "9": {
70
+ "content": "[/TOOL_RESULTS]",
71
+ "lstrip": false,
72
+ "normalized": true,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ },
77
+ "32768": {
78
+ "content": "[TOOL_RESULT]",
79
+ "lstrip": false,
80
+ "normalized": true,
81
+ "rstrip": false,
82
+ "single_word": false,
83
+ "special": true
84
+ }
85
+ },
86
+ "additional_special_tokens": [],
87
+ "bos_token": "<s>",
88
+ "chat_template": "{{bos_token}}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ ' [INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + ' ' + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
89
+ "clean_up_tokenization_spaces": false,
90
+ "eos_token": "</s>",
91
+ "legacy": true,
92
+ "model_max_length": 1000000000000000019884624838656,
93
+ "pad_token": null,
94
+ "sp_model_kwargs": {},
95
+ "spaces_between_special_tokens": false,
96
+ "tokenizer_class": "LlamaTokenizer",
97
+ "unk_token": "<unk>",
98
+ "use_default_system_prompt": false
99
+ }