Updating model files
Browse files
README.md
CHANGED
@@ -2,6 +2,17 @@
|
|
2 |
license: other
|
3 |
inference: false
|
4 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
# Vicuna 7B 1.1 GPTQ 4bit 128g
|
6 |
|
7 |
This is a 4-bit GPTQ version of the [Vicuna 7B 1.1 model](https://huggingface.co/lmsys/vicuna-7b-delta-v1.1).
|
@@ -18,7 +29,7 @@ I have the following Vicuna 1.1 repositories available:
|
|
18 |
* [Unquantized 13B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-13B-1.1-HF)
|
19 |
* [GPTQ quantized 4bit 13B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g)
|
20 |
* [GPTQ quantized 4bit 13B 1.1 for CPU - GGML format for `llama.cpp`](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g-GGML)
|
21 |
-
|
22 |
**7B models:**
|
23 |
* [Unquantized 7B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-7B-1.1-HF)
|
24 |
* [GPTQ quantized 4bit 7B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g)
|
@@ -73,6 +84,18 @@ Then link that into `text-generation-webui/repositories` as described above.
|
|
73 |
|
74 |
Or just use `vicuna-7B-1.1-GPTQ-4bit-128g.no-act-order.pt` as mentioned above.
|
75 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
76 |
# Vicuna Model Card
|
77 |
|
78 |
## Model details
|
@@ -111,4 +134,4 @@ A preliminary evaluation of the model quality is conducted by creating a set of
|
|
111 |
|
112 |
## Major updates of weights v1.1
|
113 |
- Refactor the tokenization and separator. In Vicuna v1.1, the separator has been changed from `"###"` to the EOS token `"</s>"`. This change makes it easier to determine the generation stop criteria and enables better compatibility with other libraries.
|
114 |
-
- Fix the supervised fine-tuning loss computation for better model quality.
|
|
|
2 |
license: other
|
3 |
inference: false
|
4 |
---
|
5 |
+
<div style="width: 100%;">
|
6 |
+
<img src="https://i.imgur.com/EBdldam.jpg" alt="TheBlokeAI" style="width: 100%; min-width: 400px; display: block; margin: auto;">
|
7 |
+
</div>
|
8 |
+
<div style="display: flex; justify-content: space-between; width: 100%;">
|
9 |
+
<div style="display: flex; flex-direction: column; align-items: flex-start;">
|
10 |
+
<p><a href="https://discord.gg/UBgz4VXf">Chat & support: my new Discord server</a></p>
|
11 |
+
</div>
|
12 |
+
<div style="display: flex; flex-direction: column; align-items: flex-end;">
|
13 |
+
<p><a href="https://www.patreon.com/TheBlokeAI">Want to contribute? Patreon coming soon!</a></p>
|
14 |
+
</div>
|
15 |
+
</div>
|
16 |
# Vicuna 7B 1.1 GPTQ 4bit 128g
|
17 |
|
18 |
This is a 4-bit GPTQ version of the [Vicuna 7B 1.1 model](https://huggingface.co/lmsys/vicuna-7b-delta-v1.1).
|
|
|
29 |
* [Unquantized 13B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-13B-1.1-HF)
|
30 |
* [GPTQ quantized 4bit 13B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g)
|
31 |
* [GPTQ quantized 4bit 13B 1.1 for CPU - GGML format for `llama.cpp`](https://huggingface.co/TheBloke/vicuna-13B-1.1-GPTQ-4bit-128g-GGML)
|
32 |
+
|
33 |
**7B models:**
|
34 |
* [Unquantized 7B 1.1 model for GPU - HF format](https://huggingface.co/TheBloke/vicuna-7B-1.1-HF)
|
35 |
* [GPTQ quantized 4bit 7B 1.1 for GPU - `safetensors` and `pt` formats](https://huggingface.co/TheBloke/vicuna-7B-1.1-GPTQ-4bit-128g)
|
|
|
84 |
|
85 |
Or just use `vicuna-7B-1.1-GPTQ-4bit-128g.no-act-order.pt` as mentioned above.
|
86 |
|
87 |
+
## Want to support my work?
|
88 |
+
|
89 |
+
I've had a lot of people ask if they can contribute. I love providing models and helping people, but it is starting to rack up pretty big cloud computing bills.
|
90 |
+
|
91 |
+
So if you're able and willing to contribute, it'd be most gratefully received and will help me to keep providing models, and work on various AI projects.
|
92 |
+
|
93 |
+
Donaters will get priority support on any and all AI/LLM/model questions, and I'll gladly quantise any model you'd like to try.
|
94 |
+
|
95 |
+
* Patreon: coming soon! (just awaiting approval)
|
96 |
+
* Ko-Fi: https://ko-fi.com/TheBlokeAI
|
97 |
+
* Discord: https://discord.gg/UBgz4VXf
|
98 |
+
|
99 |
# Vicuna Model Card
|
100 |
|
101 |
## Model details
|
|
|
134 |
|
135 |
## Major updates of weights v1.1
|
136 |
- Refactor the tokenization and separator. In Vicuna v1.1, the separator has been changed from `"###"` to the EOS token `"</s>"`. This change makes it easier to determine the generation stop criteria and enables better compatibility with other libraries.
|
137 |
+
- Fix the supervised fine-tuning loss computation for better model quality.
|