|
Quantization made by Richard Erkhov. |
|
|
|
[Github](https://github.com/RichardErkhov) |
|
|
|
[Discord](https://discord.gg/pvy7H8DZMG) |
|
|
|
[Request more models](https://github.com/RichardErkhov/quant_request) |
|
|
|
|
|
OPT-1.3B-SFT-DSChatLoRA - AWQ |
|
- Model creator: https://huggingface.co/kaitchup/ |
|
- Original model: https://huggingface.co/kaitchup/OPT-1.3B-SFT-DSChatLoRA/ |
|
|
|
|
|
|
|
|
|
Original model description: |
|
--- |
|
license: cc-by-nc-sa-4.0 |
|
datasets: |
|
- Dahoas/rm-static |
|
- Dahoas/synthetic-instruct-gptj-pairwise |
|
- Anthropic/hh-rlhf |
|
language: |
|
- en |
|
--- |
|
|
|
# Model Card for Model ID |
|
|
|
This a model is a fine-tuned with SFT using DeepSpeed Chat. |
|
It is based on OPT-1.3M.B |
|
|
|
## Model Details |
|
|
|
### Model Description |
|
|
|
- **Developed by:** [The Kaitchup](https://kaitchup.substack.com/) |
|
- **Model type:** Causal |
|
- **Language(s) (NLP):** English |
|
- **License:** cc-by-nc-sa-4.0 |
|
- **Finetuned from model:** [facebook/opt-1.3b](https://huggingface.co/facebook/opt-1.3b) |
|
|
|
### Model Sources |
|
|
|
The model has been trained with the procedure described in this article: |
|
|
|
[Train Instruct LLMs On Your GPU with DeepSpeed Chat — Step #1: Supervised Fine-tuning](https://kaitchup.substack.com/p/train-instruct-llms-on-your-gpu-with) |
|
|
|
|