Llama 2 7B quantized with AutoGPTQ V0.3.0.

  • Group size: 32
  • Data type: INT4

This model is compatible with the first version of QA-LoRA.

To fine-tune it with QA-LoRA, follow this tutorial: Fine-tune Quantized Llama 2 on Your GPU with QA-LoRA

Downloads last month
32
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Space using kaitchup/Llama-2-7b-4bit-32g-autogptq 1