|
--- |
|
license: apache-2.0 |
|
language: |
|
- ko |
|
library_name: transformers |
|
pipeline_tag: text-generation |
|
datasets: |
|
- maywell/ko_Ultrafeedback_binarized |
|
--- |
|
|
|
**Explanation** |
|
- With the base model, applied DPO to the small amount of layers with the open dataset , saved just the adapter part |
|
- Merged the base model and the tuned adapter together |
|
|
|
**Base Model** |
|
- [beomi/OPEN-SOLAR-KO-10.7B](https://huggingface.co/beomi/OPEN-SOLAR-KO-10.7B) |
|
|
|
**Used Corpus** |
|
- [maywell/ko_Ultrafeedback_binarized](https://huggingface.co/datasets/maywell/ko_Ultrafeedback_binarized) |
|
|
|
**Score** |
|
|Average|Ko-ARC|Ko-HellaSwag|Ko-MMLU|Ko-TruthfulQA|Ko-CommonGen V2| |
|
|:---:|:---:|:---:|:---:|:---:|:---:| |
|
|52.83|50|60.55|48.8|71.51|43.65|61.16| |
|
|
|
**Log** |
|
- 2024.01.25: Initial version Upload |
|
- 2024.02.10: Readme updated |
|
- 2024.02.11: Score updated |
|
|
|
**LICENSE** |
|
- Apache 2.0 |
|
|
|
**Citation** |
|
- beomi/OPEN-SOLAR-KO-10.7B |
|
``` |
|
@misc {solar_ko_junbum_2023, |
|
author = { {L. Junbum} }, |
|
title = { Solar-Ko-10.7b }, |
|
year = 2024, |
|
url = { https://huggingface.co/beomi/SOLAR-KO-10.7B }, |
|
publisher = { Hugging Face } |
|
} |
|
|
|
``` |