starble-dev's picture
Update README.md
51d23ea verified
|
raw
history blame
2.07 kB
metadata
license: apache-2.0
tags:
  - mistral
  - conversational
  - text-generation-inference
base_model: intervitens/mini-magnum-12b-v1.1
library_name: transformers

Sampling:
Mistral-Nemo-12B is very sensitive to the temperature sampler, try values near 0.3 at first or else you will get some weird results. This is mentioned by MistralAI at their Transformers section.

Original Model: intervitens/mini-magnum-12b-v1.1

How to Use: llama.cpp

Original Model License: Apache 2.0

Release Used: b3441

Quants

Quant Type Size
Q2_K ? GB
Q3_K_S ? GB
Q3_K_M ? GB
Q3_K_L ? GB
Q4_K_S ? GB
Q4_K_M ? GB
Q5_K_S ? GB
Q5_K_M ? GB
Q6_K ? GB
Q8_0 ? GB