Update README.md
Browse files
README.md
CHANGED
@@ -14,16 +14,13 @@ This modelcard aims to be a base template for new models. It has been generated
|
|
14 |
meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization
|
15 |
|
16 |
### Model Description
|
17 |
-
meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization
|
18 |
-
|
19 |
https://onnxruntime.ai/docs/genai/howto/install.html#directml
|
20 |
|
21 |
-
Created using ONNX Runtime GenAI's builder.py
|
22 |
-
|
23 |
https://raw.githubusercontent.com/microsoft/onnxruntime-genai/main/src/python/py/models/builder.py
|
24 |
|
25 |
-
INT4 accuracy level: FP32 (float32)
|
26 |
-
|
27 |
8-bit quantization for MoE layers
|
28 |
|
29 |
- **Developed by:** Mochamad Aris Zamroni
|
@@ -33,7 +30,6 @@ INT4 accuracy level: FP32 (float32)
|
|
33 |
- **Finetuned from model [optional]:** [More Information Needed]
|
34 |
|
35 |
### Model Sources [optional]
|
36 |
-
|
37 |
https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
|
38 |
|
39 |
- **Repository:** [More Information Needed]
|
@@ -45,33 +41,22 @@ https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
|
|
45 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
46 |
|
47 |
### Direct Use
|
48 |
-
|
49 |
This is Windows DirectML optimized model.
|
50 |
|
51 |
-
Prerequisites
|
52 |
-
|
53 |
-
1. Install Python 3.10 from Windows Store:
|
54 |
-
|
55 |
https://apps.microsoft.com/detail/9pjpw5ldxlz5?hl=en-us&gl=US
|
56 |
|
57 |
2. Open command line cmd.exe
|
58 |
|
59 |
-
3. Create python virtual environment
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
|
64 |
-
|
65 |
-
python -m venv dmlgenai
|
66 |
-
|
67 |
-
dmlgenai\Scripts\activate.bat
|
68 |
-
|
69 |
pip install onnxruntime-genai-directml
|
70 |
|
71 |
|
72 |
-
|
73 |
-
[More Information Needed]
|
74 |
-
|
75 |
## How to Get Started with the Model
|
76 |
|
77 |
Use the code below to get started with the model.
|
@@ -84,11 +69,8 @@ Use the code below to get started with the model.
|
|
84 |
|
85 |
|
86 |
#### Speeds, Sizes, Times [optional]
|
87 |
-
|
88 |
15 token/s in Radeon 780M with 8GB dedicated RAM
|
89 |
|
90 |
-
|
91 |
-
|
92 |
#### Metrics
|
93 |
|
94 |
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
@@ -109,18 +91,6 @@ Use the code below to get started with the model.
|
|
109 |
|
110 |
[More Information Needed]
|
111 |
|
112 |
-
## Environmental Impact
|
113 |
-
|
114 |
-
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
|
115 |
-
|
116 |
-
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
|
117 |
-
|
118 |
-
- **Hardware Type:** [More Information Needed]
|
119 |
-
- **Hours used:** [More Information Needed]
|
120 |
-
- **Cloud Provider:** [More Information Needed]
|
121 |
-
- **Compute Region:** [More Information Needed]
|
122 |
-
- **Carbon Emitted:** [More Information Needed]
|
123 |
-
|
124 |
## Technical Specifications [optional]
|
125 |
|
126 |
### Model Architecture and Objective
|
|
|
14 |
meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization
|
15 |
|
16 |
### Model Description
|
17 |
+
meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization<br>
|
|
|
18 |
https://onnxruntime.ai/docs/genai/howto/install.html#directml
|
19 |
|
20 |
+
Created using ONNX Runtime GenAI's builder.py<br>
|
|
|
21 |
https://raw.githubusercontent.com/microsoft/onnxruntime-genai/main/src/python/py/models/builder.py
|
22 |
|
23 |
+
INT4 accuracy level: FP32 (float32)<br>
|
|
|
24 |
8-bit quantization for MoE layers
|
25 |
|
26 |
- **Developed by:** Mochamad Aris Zamroni
|
|
|
30 |
- **Finetuned from model [optional]:** [More Information Needed]
|
31 |
|
32 |
### Model Sources [optional]
|
|
|
33 |
https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
|
34 |
|
35 |
- **Repository:** [More Information Needed]
|
|
|
41 |
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
42 |
|
43 |
### Direct Use
|
|
|
44 |
This is Windows DirectML optimized model.
|
45 |
|
46 |
+
Prerequisites:<br>
|
47 |
+
1. Install Python 3.10 from Windows Store:<br>
|
|
|
|
|
48 |
https://apps.microsoft.com/detail/9pjpw5ldxlz5?hl=en-us&gl=US
|
49 |
|
50 |
2. Open command line cmd.exe
|
51 |
|
52 |
+
3. Create python virtual environment and install onnxruntime-genai-directml<br>
|
53 |
+
mkdir c:\temp<br>
|
54 |
+
cd c:\temp<br>
|
55 |
+
python -m venv dmlgenai<br>
|
56 |
+
dmlgenai\Scripts\activate.bat<br>
|
|
|
|
|
|
|
|
|
|
|
57 |
pip install onnxruntime-genai-directml
|
58 |
|
59 |
|
|
|
|
|
|
|
60 |
## How to Get Started with the Model
|
61 |
|
62 |
Use the code below to get started with the model.
|
|
|
69 |
|
70 |
|
71 |
#### Speeds, Sizes, Times [optional]
|
|
|
72 |
15 token/s in Radeon 780M with 8GB dedicated RAM
|
73 |
|
|
|
|
|
74 |
#### Metrics
|
75 |
|
76 |
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
|
|
|
91 |
|
92 |
[More Information Needed]
|
93 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
94 |
## Technical Specifications [optional]
|
95 |
|
96 |
### Model Architecture and Objective
|