dddsaty's picture
Update README.md
313bf88 verified
---
license: apache-2.0
language:
- ko
library_name: transformers
pipeline_tag: text-generation
datasets:
- maywell/ko_Ultrafeedback_binarized
---
**Explanation**
- With the base model, applied DPO to the small amount of layers with the open dataset , saved just the adapter part
- Merged the base model and the tuned adapter together
**Base Model**
- [beomi/OPEN-SOLAR-KO-10.7B](https://huggingface.co/beomi/OPEN-SOLAR-KO-10.7B)
**Used Corpus**
- [maywell/ko_Ultrafeedback_binarized](https://huggingface.co/datasets/maywell/ko_Ultrafeedback_binarized)
**Score**
|Average|Ko-ARC|Ko-HellaSwag|Ko-MMLU|Ko-TruthfulQA|Ko-CommonGen V2|
|:---:|:---:|:---:|:---:|:---:|:---:|
|52.83|50|60.55|48.8|71.51|43.65|61.16|
**Log**
- 2024.01.25: Initial version Upload
- 2024.02.10: Readme updated
- 2024.02.11: Score updated
**LICENSE**
- Apache 2.0
**Citation**
- beomi/OPEN-SOLAR-KO-10.7B
```
@misc {solar_ko_junbum_2023,
author = { {L. Junbum} },
title = { Solar-Ko-10.7b },
year = 2024,
url = { https://huggingface.co/beomi/SOLAR-KO-10.7B },
publisher = { Hugging Face }
}
```