kroonen commited on
Commit
3ea6872
·
1 Parent(s): 207bab5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -0
README.md CHANGED
@@ -1,3 +1,29 @@
1
  ---
2
  license: mit
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
  license: mit
3
  ---
4
+
5
+ # Model description
6
+
7
+ LLAMA2-stablebeluga-Q4_0 GGML is a language model trained by Stability AI on top of Meta AI. This model is based on the original LLAMA-2, but with a couple of key changes. It has been converted to F32 before being quantized to 4 bits. These alterations make the model more efficient in terms of memory and computational requirements, without significantly compromising its language understanding and generation capabilities.
8
+
9
+ # Intended uses & limitations
10
+
11
+ ## How to use
12
+
13
+ This model can be used with llama.cpp (or similar) for a variety of natural language understanding and generation tasks. These include, but are not limited to, text completion, text generation, conversation modeling, and semantic similarity estimation.
14
+
15
+ ## Limitations and bias
16
+
17
+ While this model is designed to understand and generate human-like text, it has a few limitations:
18
+
19
+ 1. It might generate incorrect or nonsensical responses if the input prompt is ambiguous or lacks sufficient context.
20
+ 2. It is based on the data it was trained on and therefore might reflect the biases present in those data.
21
+ 3. Despite the conversion and quantization, this model might still require substantial computational resources for large-scale tasks.
22
+
23
+ # Training data
24
+
25
+ LLAMA-2-Q4_0 GGML model was trained on the same data as the original LLAMA-2 by Stability AI (Stable Beluga). For more details, please refer to the Stable Beluga 2's model card.
26
+
27
+ # Evaluations
28
+
29
+ The performance is similar to that of the original LLAMA2-stablebeluga, with a slight drop due to the quantization process. More specific evaluation results will be added as they become available.