|
--- |
|
license: apache-2.0 |
|
base_model: mistralai/Mistral-7B-v0.1 |
|
datasets: |
|
- Open-Orca/SlimOrca |
|
- HuggingFaceH4/no_robots |
|
- Intel/orca_dpo_pairs |
|
- rizerphe/glaive-function-calling-v2-zephyr |
|
- codefuse-ai/Evol-instruction-66k |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
--- |
|
|
|
# pic_7B_mistral_Full_v0.1 |
|
PIC_7B_Mistral (First phase) |
|
|
|
This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) |
|
A curated, decontaminated subset of datasets used have been mentioned in the model card. |
|
All used datasets are public as of the time of release of this model. |
|
|
|
Collaborate or Consult me - [Twitter](https://twitter.com/4evaBehindSOTA), [Discord](https://discord.gg/ftEM63pzs2) |
|
|
|
*Recommended format is ChatML, Alpaca will work but take care of EOT token* |
|
|
|
#### Chat Model Inference |
|
|
|
|
|
## Model description |
|
|
|
First generic model of Project PIC (Partner-in-Crime) in 7B range. |
|
Trying a bunch of things and seeing what sticks right now. |
|
|
|
Empathy + Coder + Instruction/json/function adherence is my game. |
|
Finding lots of challenges and insights in this effort, patience is key. |
|
![image/png](https://cdn-uploads.huggingface.co/production/uploads/6398bf222da24ee95b51c8d8/Fl-pRCsC_lvnuoP734hsJ.png) |
|
|
|
## Intended uses & limitations |
|
|
|
Should be useful in generic capacity. |
|
Demonstrates little bit of everything. |
|
|
|
Basic tests in - |
|
Roleplay: Adherence to character present. |
|
json/function-calling: Passing |
|
Coding: To be evaluated |
|
|
|
|
|
## Training procedure |
|
SFT + DPO |
|
|
|
### Training results |
|
To be evaluated |
|
|
|
### Framework versions |
|
|
|
- Transformers 4.35.2 |
|
- Pytorch 2.0.1 |
|
- Datasets 2.15.0 |
|
- Tokenizers 0.15.0 |