rogkesavan commited on
Commit
54cf961
·
verified ·
1 Parent(s): 89c2688

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -0
README.md CHANGED
@@ -69,7 +69,27 @@ print(assistant_response)
69
  ```
70
 
71
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
72
 
 
73
  ### Datasets and Fine-Tuning
74
 
75
  The following fine-tuning datasets are leveraged to enhance specific model capabilities:
 
69
  ```
70
 
71
  ---
72
+ #### Quantized Models Available for Download
73
+
74
+ | **Quantized Model Version** | **Description** |
75
+ |-------------------------------------------------------------------------------------------------------------------|---------------------------------------------------------------------------------|
76
+ | [**Nidum-Llama-3.2-3B-Uncensored-F16.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/Nidum-Llama-3.2-3B-Uncensored-F16.gguf) | Full 16-bit floating point precision for maximum accuracy on high-end GPUs. |
77
+ | [**model-Q2_K.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q2_K.gguf) | Optimized for minimal memory usage with lower precision, suitable for edge cases.|
78
+ | [**model-Q3_K_L.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q3_K_L.gguf) | Balanced precision with enhanced memory efficiency for medium-range devices. |
79
+ | [**model-Q3_K_M.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q3_K_M.gguf) | Mid-range quantization for moderate precision and memory usage balance. |
80
+ | [**model-Q3_K_S.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q3_K_S.gguf) | Smaller quantization steps, offering moderate precision with reduced memory use.|
81
+ | [**model-Q4_0_4_4.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_0_4_4.gguf) | Performance-optimized for low memory, ideal for lightweight deployment. |
82
+ | [**model-Q4_0_4_8.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_0_4_8.gguf) | Extended quantization balancing memory use and inference speed. |
83
+ | [**model-Q4_0_8_8.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_0_8_8.gguf) | Advanced memory precision targeting larger contexts. |
84
+ | [**model-Q4_K_M.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_K_M.gguf) | High-efficiency quantization for moderate GPU resources. |
85
+ | [**model-Q4_K_S.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q4_K_S.gguf) | Optimized for smaller-scale operations with compact memory footprint. |
86
+ | [**model-Q5_K_M.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q5_K_M.gguf) | Balances performance and precision, ideal for robust inferencing environments. |
87
+ | [**model-Q5_K_S.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q5_K_S.gguf) | Moderate quantization targeting performance with minimal resource usage. |
88
+ | [**model-Q6_K.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-Q6_K.gguf) | High-precision quantization for accurate and stable inferencing tasks. |
89
+ | [**model-TQ1_0.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-TQ1_0.gguf) | Experimental quantization for targeted applications in test environments. |
90
+ | [**model-TQ2_0.gguf**](https://huggingface.co/nidum/Nidum-Llama-3.2-3B-Uncensored-GGUF/blob/main/model-TQ2_0.gguf) | High-performance tuning for experimental use cases and flexible precision. |
91
 
92
+ ---
93
  ### Datasets and Fine-Tuning
94
 
95
  The following fine-tuning datasets are leveraged to enhance specific model capabilities: