File size: 3,127 Bytes
379aa4a e4a285a fb4701a 379aa4a e4a285a 379aa4a fb4701a 379aa4a e4a285a a63b5a6 e4a285a 379aa4a e4a285a 8a75dbb e4a285a 379aa4a a63b5a6 379aa4a 8a75dbb 379aa4a 8a75dbb 379aa4a 8a75dbb 379aa4a 8a75dbb 379aa4a 8a75dbb 379aa4a e4a285a 6fc6e38 e4a285a 6fc6e38 e4a285a 1ad4d5f e4a285a 1ad4d5f |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 |
---
model-index:
- name: EEVE-Instruct-Math-10.8B
results:
- task:
type: text-generation
dataset:
name: gsm8k-ko
type: gsm8k
metrics:
- name: pass@1
type: pass@1
value: 0.4845
verified: false
base_model:
- yanolja/EEVE-Korean-Instruct-10.8B-v1.0
- kuotient/EEVE-Math-10.8B-SFT
tags:
- merge
license: cc-by-sa-4.0
language:
- ko
---
# EEVE-Instruct-Math-10.8B
`EEVE-Math` ํ๋ก์ ํธ๋
- Orca-Math-200k ๋ฒ์ญ ([Orca-Math: Unlocking the potential of SLMs in Grade School Math](https://arxiv.org/pdf/2402.14830.pdf))
- gsm8k ๋ฒ์ญ, lm_eval ํ์ฉ
- Mergekit์ ์ด์ฉํ dare-ties ์ฌ์ฉ ([DARE](https://arxiv.org/abs/2311.03099))
์ ๋ํ ๋ด์ฉ์ ํฌ๊ดํ๊ณ ์์ต๋๋ค.
> ์ด ๋ชจ๋ธ์ EEVE-Math์ EEVE-Instruct์ dare-ties๋ก ๋ณํฉํ ๋ณํฉ ๋ชจ๋ธ์
๋๋ค. ์ด ํ๋ก์ ํธ๋ ์ด๋ฐ ๊ณผ์ ์ ํตํด ํนํ ๋ชจ๋ธ์ EEVE-Math์ ์ฑ๋ฅ์ ๋ง์ด ์์ง ์๊ณ Instruct ๋ชจ๋ธ์ ์ฌ์ฉ์ฑ์ ์ ์งํ ์ ์์์ ๋ณด์ฌ์ฃผ๋ Proof of concept์ ์ฑ๊ฒฉ์ ๊ฐ์ง๊ณ ์์ต๋๋ค.
| Model | gsm8k-ko(pass@1) |
|---|---|
| EEVE(Base) | 0.4049 |
| [EEVE-Math](https://huggingface.co/kuotient/EEVE-Math-10.8B) (epoch 1) | 0.508 |
| EEVE-Math (epoch 2) | **0.539** |
| [EEVE-Instruct](https://huggingface.co/yanolja/EEVE-Korean-Instruct-10.8B-v1.0) | 0.4511 |
| EEVE-Instruct + Math | **0.4845** |
## Merge Details
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [yanolja/EEVE-Korean-Instruct-10.8B-v1.0](https://huggingface.co/yanolja/EEVE-Korean-Instruct-10.8B-v1.0) as a base.
### Models Merged
The following models were included in the merge:
* [kuotient/EEVE-Math-10.8B](https://huggingface.co/kuotient/EEVE-Math-10.8B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: yanolja/EEVE-Korean-10.8B-v1.0
# no parameters necessary for base model
- model: yanolja/EEVE-Korean-Instruct-10.8B-v1.0
parameters:
density: 0.53
weight: 0.6
- model: kuotient/EEVE-Math-10.8B
parameters:
density: 0.53
weight: 0.4
merge_method: dare_ties
base_model: yanolja/EEVE-Korean-10.8B-v1.0
parameters:
int8_mask: true
dtype: bfloat16
```
## Evaluation
[gsm8k-ko](https://huggingface.co/datasets/kuotient/gsm8k-ko), kobest
```
git clone https://github.com/kuotient/lm-evaluation-harness.git
cd lm-evaluation-harness
pip install -e .
```
```
lm_eval --model hf \
--model_args pretrained=yanolja/EEVE-Korean-Instruct-2.8B-v1.0 \
--tasks gsm8k-ko \
--device cuda:0 \
--batch_size auto:4
```
| Model | gsm8k(pass@1) | boolq(acc) | copa(acc) | hellaswag(acc) | Overall |
|---|---|---|---|---|---|
| yanolja/EEVE-Korean-10.8B-v1.0 | 0.4049 | - | - | - | - | - |
| yanolja/EEVE-Korean-Instruct-10.8B-v1.0 | 0.4511 | **0.8668** | **0.7450** | 0.4940 | 0.6392 |
| [**EEVE-Math-10.8B**](https://huggingface.co/kuotient/EEVE-Math-10.8B) | **0.5390** | 0.8027 | 0.7260 | 0.4760 | 0.6359 |
| **EEVE-Instruct-Math-10.8B** | 0.4845 | 0.8519 | 0.7410 | **0.4980** | **0.6439** | |