--- license: llama2 pipeline_tag: text-generation tags: - llama - llama2 --- # Introduction Astramix is a merge of various Llama-2-7b finetune models, using [ties-merge method](https://github.com/cg123/ties-merge), thanks for [Chargoddard](https://huggingface.co/chargoddard). Subsequently, lora merge script was used, created by [zarakiquemparte](https://github.com/zarakiquemparte/zaraki-tools) ([link](https://huggingface.co/zarakiquemparte) to the Hugging Face profile). Approximate rating for this model, noticed in short-time use: * Great roleplay capabilities, limited only by model parameters. * Poor quality reasoning, because of the model limitations. * Almost censorship absence (yet some parts can be generated in the output). Feel free to test the model. Following base model used for merge: [Llama-2-7B-fp16](https://huggingface.co/TheBloke/Llama-2-7B-fp16) #### List of models used for merge: * [Nous-Hermes-llama-2-7b](https://huggingface.co/NousResearch/Nous-Hermes-llama-2-7b) * [airoboros-l2-7b-2.1](https://huggingface.co/jondurbin/airoboros-l2-7b-2.1) * [orca_mini_v3_7b ](https://huggingface.co/psmathur/orca_mini_v3_7b) * [Platypus2-7B](https://huggingface.co/garage-bAInd/Platypus2-7B) * [Tulpar-7b-v0](https://huggingface.co/HyperbeeAI/Tulpar-7b-v0) Then, two LoRAs was merged into basemix model, using script specified above: * [limarp-llama2-v2](https://huggingface.co/lemonilia/limarp-llama2-v2) (Licensed under AGPLv3) * [airoboros-lmoe-7b-2.1](https://huggingface.co/jondurbin/airoboros-lmoe-7b-2.1) (Utilizing creative version) **Here is quantized versions of the model:** * [GGUF fp16 and Q5_K_M](https://huggingface.co/androlike/astramix_l2_7b_gguf) * [GPTQ 4bit 128g](https://huggingface.co/androlike/astramix_l2_7b_4bit_128g_gptq) #### I suggest using Alpaca instruct format: ``` ### Instruction: (your instruct prompt is here) ### Response: ``` ## Limitations and risks Llama2 and its derivatives (finetunes) is licensed under LLama 2 Community License, various finetunes or (Q)LoRAs has appropriate licenses depending on used datasets in finetuning or training Low-Rank Adaptations. This mix can generate heavily biased output, which aren't suitable for minors or common audience due to using limarp in the merge.