Upload folder using huggingface_hub (#1)
Browse files- 30b5e5c13b045c46fbd60ad2106390632816ffb5ec145f67d234bac6f78e92b8 (9f5d389aab172ae23e1c49c025c3c07ead629e93)
- a094fcd31cd8f0e391a462702a6c575d619a9614c9b6778ffe33f6de21e143b9 (284991c7768e1ce722f7f29dac749a9e08bc2148)
- b618744c54f58b56d1ea85bcb9019120703839d490305f53aaf9dd797d113c11 (72a24a38df4aca0537c042f1093fd2e95030e2a9)
- 8d28d75b3c2600e6638941e9c119e0c4abc6aa29e2e4681bf05bd211d9daee95 (8280a57c905518333279903a69b6fd6e7c18cd7d)
- fcd171f06e6d2203782388fc885cd25557f8c1db27a4ca8be41478608b4a907f (656431f44cf04c504b58414c753a54164c97f67d)
- 07b8512ee6ef61b6983c60e49e586b806e64a187b98861ec2ca43142aa3c68f2 (01a02b3a212fd3c9455c799a8ed80f49db862b52)
- 3f48285d9fca47099e42bcd1de2173ac93dce25471e02b7e00c6076f0bf34890 (c95547579bebfefa82d7c27e258fca15bf1b5dcd)
- 996b0ed3e48a4610cdc9a400a9e2bc106639cb2e4f17ab4812e4601e27487234 (ba2cf9783389bf24a8a950c132473c8c0222e324)
- 1c3af2f2dfb5d27d1d79497c701336686c833044868343c3d7f0923c68c1a56b (de23430bb7eb01203467e4c79388fb6f22652a17)
- 925e141e363719ebbb1947d4dd1bbee00e7a1fbe46b85084a3cdb949822af88a (c239f70176e99d26ca25292e0eac9a8bdb0f39a7)
- 3c1afe61e6bd1662319216ae839e7c9295c54f20a8a3386f66b2e0a4d898f94a (33cd1a607edf2b9dad00dcec08917ef36c26cd9d)
- 17934129ba8e6c8dbcb73f5a7974caf8770d6696155a810b37b1750f43c4c3ca (d12f0cb5e6b981e6e571289e11372665f7e5bb05)
- 6cd7acb9b861adc6ed49c4c8b4b163f94b426f2491974ec28f9e3ae37ca4abed (6486112fb93a2e8e15a028d712b02df62ab599b4)
- a1e9d8bc16646653b55ec512d3f16a85f8ea5f5825aafd28898c81b3d1485a75 (aae0b6d4a75ee3f1354573ec22b737f2cace3645)
- f6c87821d728c164d2d312b9ff1e93aa133e1b267d102923bfcc2b2cf0d239fe (7d9eb8e49d439ab59c4cfd6952c404dd56d9e949)
- d37749578b274ff2fdb1d54e90a6253838fa183b465fe9e39bc4f75176a64657 (56a05aac91afbe012e9fe3ee16d4b9688ad919d8)
- 490a10429d135e772a8f5def6b605571dcedafa353599cd9cbcb68e4c088516e (f782f224920b4e1b92a1441709d0174f40386fa6)
- ace13f4e7e79270ae55079dea597a9877340d3829ff0015b449088de67670790 (b28ea0b7a4cc28f75368e736e92659d4e13a32c8)
- 3f3732d2a19fb798e466dc2ac94c11c7f3144ff10e541dc4297a674de9e2428a (5208487fd198661c10e1e784ed3d62a871a6faf2)
- b006dc5a95ecb397987419c04f6359192b583f232f2b4b93d99804a8b6e5f5e3 (9f7c0fcc8638be8a973b0edf8b633aa1278cb967)
- 6a3f2a88652c5bf76b13d539f91c68c3b5043e141c8d07eef193b3ec0e2e17f8 (34ac80c00d660eb67ab6b24cc0b19e2529e547a5)
- d3540a14becff2a8f2d17b819b95a429d006dbe1b7c5c78592484a587e969f19 (95df142ba9934ec2eeb1f2f064343abbd2a40e1c)
- 56ac7c1569c613e2cf8b657c4e1725f61e5d6d66810f66c5c90680bc601b887b (9993d756e9f9446a0409ea520917dc4f1d660b33)
- 0c0e95d80f99a1e9731000027937cd107d20661a87c74526069b63e5c6b94970 (b1ffad13921009285969910aaa9a181f1b9829de)
- a4fbd6c2891a266e49af0d5ba5b33077886b5ac09d21634e9d0a73ba9e379114 (c3e6779c777b52bcc6e782d5c018ea4307e78e56)
- 846d3e4cab30d4e00f6e111a18fd1f7d01c7f2c44e399441c60b96d2518fb40c (b252ab01c1f3033d9d72dc7d0ad5ad72a59aaa56)
- 26db87ae8a66e6888cf82d9c2905675a82628a98a85480f02fdc1dd30b45c364 (0a69a14457f9c62e2ba5b9b92af2be7705307c3a)
- 1b00bad3653633e508f69f0ffc5fc66327a25261b6fc951db475ae48d71e26d6 (4087922f7327f0cdf11d2c951e11a41eee5ffe89)
- ea76e0c8cd47b339d7c938dc3f5a7f8ff62936f01e1573d393964c4484cda8ca (9053fe4feea8b4c61ff3a213eadc0f3b0ff719c1)
- README.md +21 -0
- config.json +87 -0
- model-00001-of-00029.safetensors +3 -0
- model-00002-of-00029.safetensors +3 -0
- model-00003-of-00029.safetensors +3 -0
- model-00004-of-00029.safetensors +3 -0
- model-00005-of-00029.safetensors +3 -0
- model-00006-of-00029.safetensors +3 -0
- model-00007-of-00029.safetensors +3 -0
- model-00008-of-00029.safetensors +3 -0
- model-00009-of-00029.safetensors +3 -0
- model-00010-of-00029.safetensors +3 -0
- model-00011-of-00029.safetensors +3 -0
- model-00012-of-00029.safetensors +3 -0
- model-00013-of-00029.safetensors +3 -0
- model-00014-of-00029.safetensors +3 -0
- model-00015-of-00029.safetensors +3 -0
- model-00016-of-00029.safetensors +3 -0
- model-00017-of-00029.safetensors +3 -0
- model-00018-of-00029.safetensors +3 -0
- model-00019-of-00029.safetensors +3 -0
- model-00020-of-00029.safetensors +3 -0
- model-00021-of-00029.safetensors +3 -0
- model-00022-of-00029.safetensors +3 -0
- model-00023-of-00029.safetensors +3 -0
- model-00024-of-00029.safetensors +3 -0
- model-00025-of-00029.safetensors +3 -0
- model-00026-of-00029.safetensors +3 -0
- model-00027-of-00029.safetensors +3 -0
- model-00028-of-00029.safetensors +3 -0
- model-00029-of-00029.safetensors +3 -0
- model.safetensors.index.json +0 -0
- special_tokens_map.json +23 -0
- tokenizer.json +0 -0
- tokenizer_config.json +99 -0
@@ -0,0 +1,21 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- mlx
|
5 |
+
---
|
6 |
+
|
7 |
+
# mlx-community/Mixtral-8x22B-Instruct-v0.1-8bit
|
8 |
+
This model was converted to MLX format from [`mistralai/Mixtral-8x22B-Instruct-v0.1`]() using mlx-lm version **0.9.0**.
|
9 |
+
Refer to the [original model card](https://huggingface.co/mistralai/Mixtral-8x22B-Instruct-v0.1) for more details on the model.
|
10 |
+
## Use with mlx
|
11 |
+
|
12 |
+
```bash
|
13 |
+
pip install mlx-lm
|
14 |
+
```
|
15 |
+
|
16 |
+
```python
|
17 |
+
from mlx_lm import load, generate
|
18 |
+
|
19 |
+
model, tokenizer = load("mlx-community/Mixtral-8x22B-Instruct-v0.1-8bit")
|
20 |
+
response = generate(model, tokenizer, prompt="hello", verbose=True)
|
21 |
+
```
|
@@ -0,0 +1,87 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_cross_attention": false,
|
3 |
+
"architectures": [
|
4 |
+
"MixtralForCausalLM"
|
5 |
+
],
|
6 |
+
"attention_dropout": 0.0,
|
7 |
+
"bad_words_ids": null,
|
8 |
+
"begin_suppress_tokens": null,
|
9 |
+
"bos_token_id": 1,
|
10 |
+
"chunk_size_feed_forward": 0,
|
11 |
+
"cross_attention_hidden_size": null,
|
12 |
+
"decoder_start_token_id": null,
|
13 |
+
"diversity_penalty": 0.0,
|
14 |
+
"do_sample": false,
|
15 |
+
"early_stopping": false,
|
16 |
+
"encoder_no_repeat_ngram_size": 0,
|
17 |
+
"eos_token_id": 2,
|
18 |
+
"exponential_decay_length_penalty": null,
|
19 |
+
"finetuning_task": null,
|
20 |
+
"forced_bos_token_id": null,
|
21 |
+
"forced_eos_token_id": null,
|
22 |
+
"hidden_act": "silu",
|
23 |
+
"hidden_size": 6144,
|
24 |
+
"id2label": {
|
25 |
+
"0": "LABEL_0",
|
26 |
+
"1": "LABEL_1"
|
27 |
+
},
|
28 |
+
"initializer_range": 0.02,
|
29 |
+
"intermediate_size": 16384,
|
30 |
+
"is_decoder": false,
|
31 |
+
"is_encoder_decoder": false,
|
32 |
+
"label2id": {
|
33 |
+
"LABEL_0": 0,
|
34 |
+
"LABEL_1": 1
|
35 |
+
},
|
36 |
+
"length_penalty": 1.0,
|
37 |
+
"max_length": 20,
|
38 |
+
"max_position_embeddings": 65536,
|
39 |
+
"min_length": 0,
|
40 |
+
"model_type": "mixtral",
|
41 |
+
"no_repeat_ngram_size": 0,
|
42 |
+
"num_attention_heads": 48,
|
43 |
+
"num_beam_groups": 1,
|
44 |
+
"num_beams": 1,
|
45 |
+
"num_experts_per_tok": 2,
|
46 |
+
"num_hidden_layers": 56,
|
47 |
+
"num_key_value_heads": 8,
|
48 |
+
"num_local_experts": 8,
|
49 |
+
"num_return_sequences": 1,
|
50 |
+
"output_attentions": false,
|
51 |
+
"output_hidden_states": false,
|
52 |
+
"output_router_logits": false,
|
53 |
+
"output_scores": false,
|
54 |
+
"pad_token_id": null,
|
55 |
+
"prefix": null,
|
56 |
+
"problem_type": null,
|
57 |
+
"pruned_heads": {},
|
58 |
+
"quantization": {
|
59 |
+
"group_size": 64,
|
60 |
+
"bits": 8
|
61 |
+
},
|
62 |
+
"remove_invalid_values": false,
|
63 |
+
"repetition_penalty": 1.0,
|
64 |
+
"return_dict": true,
|
65 |
+
"return_dict_in_generate": false,
|
66 |
+
"rms_norm_eps": 1e-05,
|
67 |
+
"rope_theta": 1000000.0,
|
68 |
+
"router_aux_loss_coef": 0.001,
|
69 |
+
"sep_token_id": null,
|
70 |
+
"sliding_window": null,
|
71 |
+
"suppress_tokens": null,
|
72 |
+
"task_specific_params": null,
|
73 |
+
"temperature": 1.0,
|
74 |
+
"tf_legacy_loss": false,
|
75 |
+
"tie_encoder_decoder": false,
|
76 |
+
"tie_word_embeddings": false,
|
77 |
+
"tokenizer_class": null,
|
78 |
+
"top_k": 50,
|
79 |
+
"top_p": 1.0,
|
80 |
+
"torch_dtype": "bfloat16",
|
81 |
+
"torchscript": false,
|
82 |
+
"transformers_version": "4.39.3",
|
83 |
+
"typical_p": 1.0,
|
84 |
+
"use_bfloat16": false,
|
85 |
+
"use_cache": true,
|
86 |
+
"vocab_size": 32768
|
87 |
+
}
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb6744d54e379cc56c209b3562d78a06f6dfa187016c7beb2aaa9d21af4b676f
|
3 |
+
size 5296074524
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:02a60d15535b1e0b429a3a8184943ea4d31233e3e6cd6d3686bdb7e776036089
|
3 |
+
size 5321266592
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:183004cac899d86bdf0562611195604ea6a978191c43489705f293bd6096eb32
|
3 |
+
size 5321266588
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e7aed71beed50e93ae746cbdfc92928018520cc2b9f900604383152c75b5723c
|
3 |
+
size 5321266602
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1ad402be4b0c1fc43d39d762389b5a27bc975a3abf066c4509c390d6afd148e
|
3 |
+
size 5321266610
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ecdccfd56a328f20704f66aa5e77f0d3b19fcae373763d4637f54a768d86b27
|
3 |
+
size 5321266732
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9d59cc40099b308b918befd3d187d0ce0d2dcb5fed65a2c7e68efba284de9b34
|
3 |
+
size 5321266768
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bf723ad3e1a454155ee8b2214207a6028781c9accf3db4330d7c5dbfc295997b
|
3 |
+
size 5321266778
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6928dabc1f5dc8a5778a43c3ec3c15af060514b4cf14204678a47ffaffeb8378
|
3 |
+
size 5321266776
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:14b71df96bf6998e192143bfea21ca102a7b4a89ba979e09087856ac63018e1f
|
3 |
+
size 5321266766
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d810396409ab21a36b4c41b63e2c70ebd6cea86db3389ba048a6faf69e4fbbf8
|
3 |
+
size 5321266770
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fed4d33e8fb809102851e92de6e761cc9b5eefb29c3718a2f59231d3c4d4a04c
|
3 |
+
size 5321266754
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:76f06b9b4802ea1af03b44c87cf8e87371a60a3045ebc66f3ff212c41062b0a8
|
3 |
+
size 5321266720
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:273496bc6c71c9005970ad68354be366123a4e8a85776239743238cb99cb418f
|
3 |
+
size 5321266784
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b74fb40ae399a6171237887aefc40ee93bd5dbfad4d54d756809249cd285c5c3
|
3 |
+
size 5321266782
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b433338d423e6e0dffb226d62e70739fb0e2f1bdcf6f76546eb5ee3288e5c03e
|
3 |
+
size 5321266766
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7ec1635af29bd9957455cf590f393a0d4db3703e3a3cb127811c6f3d4fc8e81f
|
3 |
+
size 5321266764
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e81c6e40ea718be515c647480e34f5daec9634b8197b8bd9f2debfa09f54e5c
|
3 |
+
size 5321266778
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d89d44e15e66e6fce9bba7240015397263ffdcecd0da8a46ee173db5db11f130
|
3 |
+
size 5321266772
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9213616e2342b9a47085640b46e0aefee32d7a9e9e5eea840ef7815ecfde73af
|
3 |
+
size 5321266778
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c269b5bafecd7f256fe33ba919cc7de99ab33a9d1fd6bdd666d37adb661e3e5b
|
3 |
+
size 5321266750
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:28902e02f9ac9b5655d068958c6ac7eecc7804d5c5b79fe98887da85b3944b9a
|
3 |
+
size 5321266764
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0fdcf31e0d42326a6868a1737c2063f5238d5a5f6a56a32de24fb12d16a2c4c9
|
3 |
+
size 5321266772
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4f3ebfc406d7351de656ba584fe5549a1d6cbd8eb7b7ea8f97f33224a4c858cf
|
3 |
+
size 5321266754
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7d0722fcb68d479f88df43092ac2a80b47cdc7cd00448b4645909ae3dc9c594b
|
3 |
+
size 5321266772
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3192abd43305e47713862e37136b7e7048884a56f5a43f4ab7b116174eaa46ed
|
3 |
+
size 5321266776
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b857e5ad08617d906024a8763daf2dc8d0ae1d47f5975a9b294b844607476f6e
|
3 |
+
size 5321266774
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c91d80c1c9d44118eb7f15fd579e9c987b7582deb18f24a5a3b8d2b790732ab2
|
3 |
+
size 5321266768
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3727875b033b34db156ac0b1aacfce305343e6007c8f6fd82376e9e90e08722a
|
3 |
+
size 641767520
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,23 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"bos_token": {
|
3 |
+
"content": "<s>",
|
4 |
+
"lstrip": false,
|
5 |
+
"normalized": true,
|
6 |
+
"rstrip": false,
|
7 |
+
"single_word": false
|
8 |
+
},
|
9 |
+
"eos_token": {
|
10 |
+
"content": "</s>",
|
11 |
+
"lstrip": false,
|
12 |
+
"normalized": true,
|
13 |
+
"rstrip": false,
|
14 |
+
"single_word": false
|
15 |
+
},
|
16 |
+
"unk_token": {
|
17 |
+
"content": "<unk>",
|
18 |
+
"lstrip": false,
|
19 |
+
"normalized": true,
|
20 |
+
"rstrip": false,
|
21 |
+
"single_word": false
|
22 |
+
}
|
23 |
+
}
|
The diff for this file is too large to render.
See raw diff
|
|
@@ -0,0 +1,99 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"added_tokens_decoder": {
|
5 |
+
"0": {
|
6 |
+
"content": "<unk>",
|
7 |
+
"lstrip": false,
|
8 |
+
"normalized": true,
|
9 |
+
"rstrip": false,
|
10 |
+
"single_word": false,
|
11 |
+
"special": true
|
12 |
+
},
|
13 |
+
"1": {
|
14 |
+
"content": "<s>",
|
15 |
+
"lstrip": false,
|
16 |
+
"normalized": true,
|
17 |
+
"rstrip": false,
|
18 |
+
"single_word": false,
|
19 |
+
"special": true
|
20 |
+
},
|
21 |
+
"2": {
|
22 |
+
"content": "</s>",
|
23 |
+
"lstrip": false,
|
24 |
+
"normalized": true,
|
25 |
+
"rstrip": false,
|
26 |
+
"single_word": false,
|
27 |
+
"special": true
|
28 |
+
},
|
29 |
+
"3": {
|
30 |
+
"content": "[INST]",
|
31 |
+
"lstrip": false,
|
32 |
+
"normalized": true,
|
33 |
+
"rstrip": false,
|
34 |
+
"single_word": false,
|
35 |
+
"special": true
|
36 |
+
},
|
37 |
+
"4": {
|
38 |
+
"content": "[/INST]",
|
39 |
+
"lstrip": false,
|
40 |
+
"normalized": true,
|
41 |
+
"rstrip": false,
|
42 |
+
"single_word": false,
|
43 |
+
"special": true
|
44 |
+
},
|
45 |
+
"5": {
|
46 |
+
"content": "[TOOL_CALLS]",
|
47 |
+
"lstrip": false,
|
48 |
+
"normalized": true,
|
49 |
+
"rstrip": false,
|
50 |
+
"single_word": false,
|
51 |
+
"special": true
|
52 |
+
},
|
53 |
+
"6": {
|
54 |
+
"content": "[AVAILABLE_TOOLS]",
|
55 |
+
"lstrip": false,
|
56 |
+
"normalized": true,
|
57 |
+
"rstrip": false,
|
58 |
+
"single_word": false,
|
59 |
+
"special": true
|
60 |
+
},
|
61 |
+
"7": {
|
62 |
+
"content": "[/AVAILABLE_TOOLS]",
|
63 |
+
"lstrip": false,
|
64 |
+
"normalized": true,
|
65 |
+
"rstrip": false,
|
66 |
+
"single_word": false,
|
67 |
+
"special": true
|
68 |
+
},
|
69 |
+
"9": {
|
70 |
+
"content": "[/TOOL_RESULTS]",
|
71 |
+
"lstrip": false,
|
72 |
+
"normalized": true,
|
73 |
+
"rstrip": false,
|
74 |
+
"single_word": false,
|
75 |
+
"special": true
|
76 |
+
},
|
77 |
+
"32768": {
|
78 |
+
"content": "[TOOL_RESULT]",
|
79 |
+
"lstrip": false,
|
80 |
+
"normalized": true,
|
81 |
+
"rstrip": false,
|
82 |
+
"single_word": false,
|
83 |
+
"special": true
|
84 |
+
}
|
85 |
+
},
|
86 |
+
"additional_special_tokens": [],
|
87 |
+
"bos_token": "<s>",
|
88 |
+
"chat_template": "{{bos_token}}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ ' [INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ ' ' + message['content'] + ' ' + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
|
89 |
+
"clean_up_tokenization_spaces": false,
|
90 |
+
"eos_token": "</s>",
|
91 |
+
"legacy": true,
|
92 |
+
"model_max_length": 1000000000000000019884624838656,
|
93 |
+
"pad_token": null,
|
94 |
+
"sp_model_kwargs": {},
|
95 |
+
"spaces_between_special_tokens": false,
|
96 |
+
"tokenizer_class": "LlamaTokenizer",
|
97 |
+
"unk_token": "<unk>",
|
98 |
+
"use_default_system_prompt": false
|
99 |
+
}
|