Update README.md
Browse files
README.md
CHANGED
@@ -73,7 +73,7 @@ GGML files are for CPU + GPU inference using [llama.cpp](https://github.com/gger
|
|
73 |
## Repositories available
|
74 |
|
75 |
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/minotaur-15B-GPTQ)
|
76 |
-
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU
|
77 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/openaccess-ai-collective/minotaur-15b)
|
78 |
|
79 |
## A note regarding context length
|
|
|
73 |
## Repositories available
|
74 |
|
75 |
* [4-bit GPTQ models for GPU inference](https://huggingface.co/TheBloke/minotaur-15B-GPTQ)
|
76 |
+
* [2, 3, 4, 5, 6 and 8-bit GGML models for CPU inference](https://huggingface.co/TheBloke/minotaur-15B-GGML)
|
77 |
* [Unquantised fp16 model in pytorch format, for GPU inference and for further conversions](https://huggingface.co/openaccess-ai-collective/minotaur-15b)
|
78 |
|
79 |
## A note regarding context length
|