ngxson HF staff Xenova HF staff commited on
Commit
0eb811a
·
verified ·
1 Parent(s): 06d923b

Upload folder using huggingface_hub (#1)

Browse files

- Upload folder using huggingface_hub (3e0a36cc370ba3a5a32c0c04fc2fc9f60f89a523)
- Upload folder using huggingface_hub (95ccdbf0c39674b400c3152e8a11df188b5976fd)


Co-authored-by: Joshua <[email protected]>

.gitattributes CHANGED
@@ -34,3 +34,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  tokenizer.json filter=lfs diff=lfs merge=lfs -text
37
+ onnx/model.onnx_data filter=lfs diff=lfs merge=lfs -text
38
+ onnx/model_fp16.onnx_data filter=lfs diff=lfs merge=lfs -text
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "ngxson/MiniThinky-1B-Llama-3.2",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
@@ -36,11 +36,11 @@
36
  "torch_dtype": "float32",
37
  "transformers.js_config": {
38
  "kv_cache_dtype": {
39
- "fp16": "float16",
40
- "q4f16": "float16"
41
  }
42
  },
43
- "transformers_version": "4.47.1",
44
  "use_cache": true,
45
  "vocab_size": 128256
46
- }
 
1
  {
2
+ "_name_or_path": "ngxson/MiniThinky-v2-1B-Llama-3.2",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
 
36
  "torch_dtype": "float32",
37
  "transformers.js_config": {
38
  "kv_cache_dtype": {
39
+ "q4f16": "float16",
40
+ "fp16": "float16"
41
  }
42
  },
43
+ "transformers_version": "4.46.3",
44
  "use_cache": true,
45
  "vocab_size": 128256
46
+ }
generation_config.json CHANGED
@@ -8,5 +8,5 @@
8
  ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
- "transformers_version": "4.47.1"
12
  }
 
8
  ],
9
  "temperature": 0.6,
10
  "top_p": 0.9,
11
+ "transformers_version": "4.46.3"
12
  }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5979818b25758b4eb68a68c1019d19b7fefcdfb6c5d0c9b4f40b9bb7c6a0d1f8
3
+ size 291652
onnx/model.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da1afbbd051a785ce18222ba53e02b79a1764a6923cdab2a46ceb0e3c993da41
3
+ size 4976812032
onnx/model_bnb4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2edd5e43ff1a83c5ef49dc6b585ff587db9064498ce591d2105e102ab425938e
3
+ size 1632159815
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:168c8d58f64d122af7061a69dcada5465ec3af63b866aaf19b25f47c834bed71
3
+ size 398882714
onnx/model_fp16.onnx_data ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c7717e320e4f5fccf344cdc8956dcf66aede9826f8f34b92b4ac300578d1346
3
+ size 2089811968
onnx/model_int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85d79af17aa267a2d5e7c870944b942cfe32b8200d2d896f3597c3ddb9a3d3a2
3
+ size 1269982484
onnx/model_q4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24eceb73708302ba138fbbc8313d0207789f7881db89d2c67c612650b7682080
3
+ size 1692976359
onnx/model_q4f16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f9ea154a73edc51c78c09f4ee299807365481d141d0bc1c3daf6b26092ffc64
3
+ size 1089909914
onnx/model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16c07caf25d766351ba5cd35ed5970ff141b18efafdfa4bb601cf28faf5c3d2e
3
+ size 1269982544
onnx/model_uint8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16c07caf25d766351ba5cd35ed5970ff141b18efafdfa4bb601cf28faf5c3d2e
3
+ size 1269982544
special_tokens_map.json CHANGED
@@ -13,5 +13,11 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": "<|finetune_right_pad_id|>"
 
 
 
 
 
 
17
  }
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "pad_token": {
17
+ "content": "<|finetune_right_pad_id|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:eb9b85b70a1b34e497c3476c315b656613e7ed0b2db8c6cf62ec83d381c98737
3
- size 17209986
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c02b5c0c3fbb4b776c5ad9ede41ef8421ce0a8be84fa0bf5812fbbc80557e21
3
+ size 11574684
tokenizer_config.json CHANGED
@@ -2054,11 +2054,15 @@
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "extra_special_tokens": {},
 
2057
  "model_input_names": [
2058
  "input_ids",
2059
  "attention_mask"
2060
  ],
2061
  "model_max_length": 131072,
2062
  "pad_token": "<|finetune_right_pad_id|>",
2063
- "tokenizer_class": "PreTrainedTokenizerFast"
 
 
 
2064
  }
 
2054
  "clean_up_tokenization_spaces": true,
2055
  "eos_token": "<|eot_id|>",
2056
  "extra_special_tokens": {},
2057
+ "max_length": 32000,
2058
  "model_input_names": [
2059
  "input_ids",
2060
  "attention_mask"
2061
  ],
2062
  "model_max_length": 131072,
2063
  "pad_token": "<|finetune_right_pad_id|>",
2064
+ "stride": 0,
2065
+ "tokenizer_class": "PreTrainedTokenizerFast",
2066
+ "truncation_side": "right",
2067
+ "truncation_strategy": "longest_first"
2068
  }