File size: 821 Bytes
7dac834 5ebd31b bc769af fdb9a0a 59729a1 2e499ac 59729a1 23d2bb9 59729a1 bc769af f56be9a bc769af d6ab674 59729a1 1bee25d b03f230 d625ec5 bf10ea8 59729a1 bf10ea8 1eaa66d 5ebd31b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 |
---
license: apache-2.0
datasets:
- argilla/ultrafeedback-binarized-preferences-cleaned
language:
- en
base_model:
- mistralai/Mistral-7B-v0.1
library_name: transformers
tags:
- transformers
- ORPO
- RLHF
- notus
- argilla
---
# Model Overview
# ππ¨πππ₯ πππ¦π:ElEmperador
![image/png](https://cdn-uploads.huggingface.co/production/uploads/64e8ea3892d9db9a93580fe3/gkDcpIxRCjBlmknN_jzWN.png)
## Model Description:
ElEmperador is an ORPO-based finetune derived from the Mistral-7B-v0.1 base model.
## Evals:
BLEU:0.209
## Results
Firstly,ORPO is a viable RLHF algorithm to improve the performance of your models along with SFT finetuning.Secondly, it also helps in aligning the modelβs outputs more closely with human preferences,
leading to more user-friendly and acceptable results. |