|
--- |
|
license: llama2 |
|
pipeline_tag: text-generation |
|
tags: |
|
- llama |
|
- llama2 |
|
--- |
|
# Introduction |
|
|
|
Astramix is a merge of various Llama-2-7b finetune models, using [ties-merge method](https://github.com/cg123/ties-merge), thanks for [Chargoddard](https://huggingface.co/chargoddard). |
|
Subsequently, lora merge script was used, created by [zarakiquemparte](https://github.com/zarakiquemparte/zaraki-tools) ([link](https://huggingface.co/zarakiquemparte) to the Hugging Face profile). |
|
|
|
Approximate rating for this model, noticed in short-time use: |
|
* Great roleplay capabilities, limited only by model parameters. |
|
* Poor quality reasoning, because of the model limitations. |
|
* Almost censorship absence (yet some parts can be generated in the output). |
|
|
|
Feel free to test the model. |
|
|
|
Following base model used for merge: [Llama-2-7B-fp16](https://huggingface.co/TheBloke/Llama-2-7B-fp16) |
|
|
|
#### List of models used for merge: |
|
* [Nous-Hermes-llama-2-7b](https://huggingface.co/NousResearch/Nous-Hermes-llama-2-7b) |
|
* [airoboros-l2-7b-2.1](https://huggingface.co/jondurbin/airoboros-l2-7b-2.1) |
|
* [orca_mini_v3_7b ](https://huggingface.co/psmathur/orca_mini_v3_7b) |
|
* [Platypus2-7B](https://huggingface.co/garage-bAInd/Platypus2-7B) |
|
* [Tulpar-7b-v0](https://huggingface.co/HyperbeeAI/Tulpar-7b-v0) |
|
|
|
Then, two LoRAs was merged into basemix model, using script specified above: |
|
* [limarp-llama2-v2](https://huggingface.co/lemonilia/limarp-llama2-v2) (Licensed under AGPLv3) |
|
* [airoboros-lmoe-7b-2.1](https://huggingface.co/jondurbin/airoboros-lmoe-7b-2.1) (Utilizing creative version) |
|
|
|
**Here is quantized versions of the model:** |
|
* [GGUF fp16 and Q5_K_M](https://huggingface.co/androlike/astramix_l2_7b_gguf) |
|
* [GPTQ 4bit 128g](https://huggingface.co/androlike/astramix_l2_7b_4bit_128g_gptq) |
|
|
|
#### I suggest using Alpaca instruct format: |
|
``` |
|
### Instruction: |
|
(your instruct prompt is here) |
|
### Response: |
|
``` |
|
## Limitations and risks |
|
Llama2 and its derivatives (finetunes) is licensed under LLama 2 Community License, various finetunes or (Q)LoRAs has appropriate licenses depending on used datasets in finetuning or training Low-Rank Adaptations. |
|
This mix can generate heavily biased output, which aren't suitable for minors or common audience due to using limarp in the merge. |