haoyang-amd commited on
Commit
39a18ee
·
verified ·
1 Parent(s): 4e3d382

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -13
README.md CHANGED
@@ -1,9 +1,7 @@
1
  ---
2
- base_model:
3
- - microsoft/Phi-3-mini-128k-instruct
4
- license: mit
5
  ---
6
- # Phi-3-mini-128k-instruct-Weight-INT4-Per-Group-AWQ-Bfloat16
7
  - ## Introduction
8
  This model was created by applying [Quark](https://quark.docs.amd.com/latest/index.html) with calibration samples from Pile dataset.
9
  - ## Quantization Stragegy
@@ -14,7 +12,7 @@ license: mit
14
  1. [Download and install Quark](https://quark.docs.amd.com/latest/install.html)
15
  2. Run the quantization script in the example folder using the following command line:
16
  ```sh
17
- export MODEL_DIR = [local model checkpoint folder] or microsoft/Phi-3-mini-128k-instruct
18
  # single GPU
19
  python3 quantize_quark.py --model_dir $MODEL_DIR \
20
  --data_type bfloat16 \
@@ -23,7 +21,7 @@ license: mit
23
  --quant_algo awq \
24
  --dataset pileval_for_awq_benchmark \
25
  --seq_len 512 \
26
- --output_dir Phi-3-mini-128k-instruct-W_Int4-Per_Group-AWQ-BFloat16 \
27
  --model_export quark_safetensors
28
  # cpu
29
  python3 quantize_quark.py --model_dir $MODEL_DIR \
@@ -33,7 +31,7 @@ license: mit
33
  --quant_algo awq \
34
  --dataset pileval_for_awq_benchmark \
35
  --seq_len 512 \
36
- --output_dir Phi-3-mini-128k-instruct-W_Int4-Per_Group-AWQ-BFloat16 \
37
  --model_export quark_safetensors \
38
  --device cpu
39
  ```
@@ -47,23 +45,25 @@ The quantization evaluation results are conducted in pseudo-quantization mode, w
47
  <tr>
48
  <td><strong>Benchmark</strong>
49
  </td>
50
- <td><strong>Phi-3-mini-4k-instruct(Bfloat16) </strong>
51
  </td>
52
- <td><strong>Phi-3-mini-4k-instruct-Weight-INT4-Per-Group-AWQ-Bfloat16(this model)</strong>
53
  </td>
54
  </tr>
55
  <tr>
56
  <td>Perplexity-wikitext2
57
  </td>
58
- <td>6.2359
59
  </td>
60
- <td>6.8193
61
  </td>
62
  </tr>
 
63
  </table>
64
 
65
  #### License
66
- Modifications copyright(c) 2024 Advanced Micro Devices, Inc. All rights reserved.
 
67
 
68
  Licensed under the Apache License, Version 2.0 (the "License");
69
  you may not use this file except in compliance with the License.
@@ -75,4 +75,4 @@ Unless required by applicable law or agreed to in writing, software
75
  distributed under the License is distributed on an "AS IS" BASIS,
76
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
77
  See the License for the specific language governing permissions and
78
- limitations under the License.
 
1
  ---
2
+ base_model: mistralai/Mistral-7B-v0.1
 
 
3
  ---
4
+ # Mistral-7B-v0.1-AWQ-G128-INT4-SYM-BF16
5
  - ## Introduction
6
  This model was created by applying [Quark](https://quark.docs.amd.com/latest/index.html) with calibration samples from Pile dataset.
7
  - ## Quantization Stragegy
 
12
  1. [Download and install Quark](https://quark.docs.amd.com/latest/install.html)
13
  2. Run the quantization script in the example folder using the following command line:
14
  ```sh
15
+ export MODEL_DIR = [local model checkpoint folder] or mistralai/Mistral-7B-v0.1
16
  # single GPU
17
  python3 quantize_quark.py --model_dir $MODEL_DIR \
18
  --data_type bfloat16 \
 
21
  --quant_algo awq \
22
  --dataset pileval_for_awq_benchmark \
23
  --seq_len 512 \
24
+ --output_dir Mistral-7B-v0.1-AWQ-G128-INT4-SYM-BF16 \
25
  --model_export quark_safetensors
26
  # cpu
27
  python3 quantize_quark.py --model_dir $MODEL_DIR \
 
31
  --quant_algo awq \
32
  --dataset pileval_for_awq_benchmark \
33
  --seq_len 512 \
34
+ --output_dir Mistral-7B-v0.1-AWQ-G128-INT4-SYM-BF16 \
35
  --model_export quark_safetensors \
36
  --device cpu
37
  ```
 
45
  <tr>
46
  <td><strong>Benchmark</strong>
47
  </td>
48
+ <td><strong>Mistral-7B-v0.1(Bfloat16) </strong>
49
  </td>
50
+ <td><strong>Mistral-7B-v0.1-AWQ-G128-INT4-SYM-BF16(this model)</strong>
51
  </td>
52
  </tr>
53
  <tr>
54
  <td>Perplexity-wikitext2
55
  </td>
56
+ <td>5.2527
57
  </td>
58
+ <td>5.4250
59
  </td>
60
  </tr>
61
+
62
  </table>
63
 
64
  #### License
65
+
66
+ Modifications copyright(c) 2024 Advanced Micro Devices,Inc. All rights reserved.
67
 
68
  Licensed under the Apache License, Version 2.0 (the "License");
69
  you may not use this file except in compliance with the License.
 
75
  distributed under the License is distributed on an "AS IS" BASIS,
76
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
77
  See the License for the specific language governing permissions and
78
+ limitations under the License.