zamroni111 commited on
Commit
bc78057
·
verified ·
1 Parent(s): 555883c

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -40
README.md CHANGED
@@ -14,16 +14,13 @@ This modelcard aims to be a base template for new models. It has been generated
14
  meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization
15
 
16
  ### Model Description
17
- meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization
18
-
19
  https://onnxruntime.ai/docs/genai/howto/install.html#directml
20
 
21
- Created using ONNX Runtime GenAI's builder.py
22
-
23
  https://raw.githubusercontent.com/microsoft/onnxruntime-genai/main/src/python/py/models/builder.py
24
 
25
- INT4 accuracy level: FP32 (float32)
26
-
27
  8-bit quantization for MoE layers
28
 
29
  - **Developed by:** Mochamad Aris Zamroni
@@ -33,7 +30,6 @@ INT4 accuracy level: FP32 (float32)
33
  - **Finetuned from model [optional]:** [More Information Needed]
34
 
35
  ### Model Sources [optional]
36
-
37
  https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
38
 
39
  - **Repository:** [More Information Needed]
@@ -45,33 +41,22 @@ https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
45
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
46
 
47
  ### Direct Use
48
-
49
  This is Windows DirectML optimized model.
50
 
51
- Prerequisites:
52
-
53
- 1. Install Python 3.10 from Windows Store:
54
-
55
  https://apps.microsoft.com/detail/9pjpw5ldxlz5?hl=en-us&gl=US
56
 
57
  2. Open command line cmd.exe
58
 
59
- 3. Create python virtual environment:
60
-
61
- mkdir c:\temp
62
-
63
- cd c:\temp
64
-
65
- python -m venv dmlgenai
66
-
67
- dmlgenai\Scripts\activate.bat
68
-
69
  pip install onnxruntime-genai-directml
70
 
71
 
72
-
73
- [More Information Needed]
74
-
75
  ## How to Get Started with the Model
76
 
77
  Use the code below to get started with the model.
@@ -84,11 +69,8 @@ Use the code below to get started with the model.
84
 
85
 
86
  #### Speeds, Sizes, Times [optional]
87
-
88
  15 token/s in Radeon 780M with 8GB dedicated RAM
89
 
90
-
91
-
92
  #### Metrics
93
 
94
  <!-- These are the evaluation metrics being used, ideally with a description of why. -->
@@ -109,18 +91,6 @@ Use the code below to get started with the model.
109
 
110
  [More Information Needed]
111
 
112
- ## Environmental Impact
113
-
114
- <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
115
-
116
- Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
117
-
118
- - **Hardware Type:** [More Information Needed]
119
- - **Hours used:** [More Information Needed]
120
- - **Cloud Provider:** [More Information Needed]
121
- - **Compute Region:** [More Information Needed]
122
- - **Carbon Emitted:** [More Information Needed]
123
-
124
  ## Technical Specifications [optional]
125
 
126
  ### Model Architecture and Objective
 
14
  meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization
15
 
16
  ### Model Description
17
+ meta-llama/Meta-Llama-3.1-8B-Instruct quantized to ONNX GenAI INT4 with Microsoft DirectML optimization<br>
 
18
  https://onnxruntime.ai/docs/genai/howto/install.html#directml
19
 
20
+ Created using ONNX Runtime GenAI's builder.py<br>
 
21
  https://raw.githubusercontent.com/microsoft/onnxruntime-genai/main/src/python/py/models/builder.py
22
 
23
+ INT4 accuracy level: FP32 (float32)<br>
 
24
  8-bit quantization for MoE layers
25
 
26
  - **Developed by:** Mochamad Aris Zamroni
 
30
  - **Finetuned from model [optional]:** [More Information Needed]
31
 
32
  ### Model Sources [optional]
 
33
  https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct
34
 
35
  - **Repository:** [More Information Needed]
 
41
  <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
42
 
43
  ### Direct Use
 
44
  This is Windows DirectML optimized model.
45
 
46
+ Prerequisites:<br>
47
+ 1. Install Python 3.10 from Windows Store:<br>
 
 
48
  https://apps.microsoft.com/detail/9pjpw5ldxlz5?hl=en-us&gl=US
49
 
50
  2. Open command line cmd.exe
51
 
52
+ 3. Create python virtual environment and install onnxruntime-genai-directml<br>
53
+ mkdir c:\temp<br>
54
+ cd c:\temp<br>
55
+ python -m venv dmlgenai<br>
56
+ dmlgenai\Scripts\activate.bat<br>
 
 
 
 
 
57
  pip install onnxruntime-genai-directml
58
 
59
 
 
 
 
60
  ## How to Get Started with the Model
61
 
62
  Use the code below to get started with the model.
 
69
 
70
 
71
  #### Speeds, Sizes, Times [optional]
 
72
  15 token/s in Radeon 780M with 8GB dedicated RAM
73
 
 
 
74
  #### Metrics
75
 
76
  <!-- These are the evaluation metrics being used, ideally with a description of why. -->
 
91
 
92
  [More Information Needed]
93
 
 
 
 
 
 
 
 
 
 
 
 
 
94
  ## Technical Specifications [optional]
95
 
96
  ### Model Architecture and Objective