Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,29 @@
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
---
|
4 |
+
|
5 |
+
# Model description
|
6 |
+
|
7 |
+
LLAMA2-stablebeluga-Q4_0 GGML is a language model trained by Stability AI on top of Meta AI. This model is based on the original LLAMA-2, but with a couple of key changes. It has been converted to F32 before being quantized to 4 bits. These alterations make the model more efficient in terms of memory and computational requirements, without significantly compromising its language understanding and generation capabilities.
|
8 |
+
|
9 |
+
# Intended uses & limitations
|
10 |
+
|
11 |
+
## How to use
|
12 |
+
|
13 |
+
This model can be used with llama.cpp (or similar) for a variety of natural language understanding and generation tasks. These include, but are not limited to, text completion, text generation, conversation modeling, and semantic similarity estimation.
|
14 |
+
|
15 |
+
## Limitations and bias
|
16 |
+
|
17 |
+
While this model is designed to understand and generate human-like text, it has a few limitations:
|
18 |
+
|
19 |
+
1. It might generate incorrect or nonsensical responses if the input prompt is ambiguous or lacks sufficient context.
|
20 |
+
2. It is based on the data it was trained on and therefore might reflect the biases present in those data.
|
21 |
+
3. Despite the conversion and quantization, this model might still require substantial computational resources for large-scale tasks.
|
22 |
+
|
23 |
+
# Training data
|
24 |
+
|
25 |
+
LLAMA-2-Q4_0 GGML model was trained on the same data as the original LLAMA-2 by Stability AI (Stable Beluga). For more details, please refer to the Stable Beluga 2's model card.
|
26 |
+
|
27 |
+
# Evaluations
|
28 |
+
|
29 |
+
The performance is similar to that of the original LLAMA2-stablebeluga, with a slight drop due to the quantization process. More specific evaluation results will be added as they become available.
|