EricB HF staff commited on
Commit
064911d
·
verified ·
1 Parent(s): 51959b4

Upload model

Browse files
.gitattributes CHANGED
@@ -33,3 +33,10 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ mistral0.3-7b-instruct-f8e4m3.uqff filter=lfs diff=lfs merge=lfs -text
37
+ mistral0.3-7b-instruct-hqq4.uqff filter=lfs diff=lfs merge=lfs -text
38
+ mistral0.3-7b-instruct-hqq8.uqff filter=lfs diff=lfs merge=lfs -text
39
+ mistral0.3-7b-instruct-q3k.uqff filter=lfs diff=lfs merge=lfs -text
40
+ mistral0.3-7b-instruct-q4k.uqff filter=lfs diff=lfs merge=lfs -text
41
+ mistral0.3-7b-instruct-q5k.uqff filter=lfs diff=lfs merge=lfs -text
42
+ mistral0.3-7b-instruct-q8_0.uqff filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "MistralForCausalLM"
4
+ ],
5
+ "attention_dropout": 0.0,
6
+ "bos_token_id": 1,
7
+ "eos_token_id": 2,
8
+ "hidden_act": "silu",
9
+ "hidden_size": 4096,
10
+ "initializer_range": 0.02,
11
+ "intermediate_size": 14336,
12
+ "max_position_embeddings": 32768,
13
+ "model_type": "mistral",
14
+ "num_attention_heads": 32,
15
+ "num_hidden_layers": 32,
16
+ "num_key_value_heads": 8,
17
+ "rms_norm_eps": 1e-05,
18
+ "rope_theta": 1000000.0,
19
+ "sliding_window": null,
20
+ "tie_word_embeddings": false,
21
+ "torch_dtype": "bfloat16",
22
+ "transformers_version": "4.42.0.dev0",
23
+ "use_cache": true,
24
+ "vocab_size": 32768
25
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.42.0.dev0"
6
+ }
mistral0.3-7b-instruct-f8e4m3.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e9eb637c638d005ba42a940154cb20ab466da56d526d0f3403ee1ad2182d0c54
3
+ size 7113566578
mistral0.3-7b-instruct-hqq4.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6be3d2aa28415a4d4cd9cd573ae5bb9c78ff6963f900b59645c29769fe6f8e59
3
+ size 4445999994
mistral0.3-7b-instruct-hqq8.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc6a38ec801f904bec1a71123d3b47e6e2e73e0cc630ad3feef08345b5f63388
3
+ size 8002769834
mistral0.3-7b-instruct-q3k.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:05242e90cd2bc37db7c3a067603bdf00d8549511b278c12892a7f24114ce8f66
3
+ size 3056622274
mistral0.3-7b-instruct-q4k.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f49603285cc6a24942511296cba0151a8d9c81073fe97845a3464a4da0d84fa
3
+ size 4001389290
mistral0.3-7b-instruct-q5k.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e5c55e4e77ead0deef64e227f0c5cf81a6058769c3b30b76d1917aa08139cdf
3
+ size 4890581826
mistral0.3-7b-instruct-q8_0.uqff ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5b9cc8730409394a09edddcec38fc47b9caa1264874a7c200f75cbbee78c75e0
3
+ size 7558159202
residual.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff71024190bb6f21d57e28eefb6acfa6d31ab2311f86430b6dfb285b8f165943
3
+ size 268975440
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff