File size: 1,259 Bytes
395f5e4 7623537 395f5e4 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 |
---
language:
- en
- ko
license: other
license_name: exaone
license_link: LICENSE
tags:
- text-generation-inference
- transformers
- trl
- sft
- reasoning
- lg-ai
- exaone
- exaone-3.5
- o1
base_model: LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct
datasets:
- KingNish/reasoning-base-20k
---
# Model Description
An uncensored reasoning EXAONE 3.5 model trained on reasoning data. Now with a full epoch!
It has been trained using improved training code, and gives an improved performance.
Here is what inference code you should use:
```py
# DEBUGGING IN PROGRESS, check later
```
- **Trained by:** [Piotr Zalewski](https://huggingface.co/lunahr)
- **License:** exaone
- **Finetuned from model:** [LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct](https://huggingface.co/LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct)
- **Dataset used:** [KingNish/reasoning-base-20k](https://huggingface.co/datasets/KingNish/reasoning-base-20k)
This Llama model was trained faster than [Unsloth](https://github.com/unslothai/unsloth) using [custom training code](https://www.kaggle.com/code/piotr25691/distributed-hf-training-with-2xt4).
Visit https://www.kaggle.com/code/piotr25691/distributed-hf-training-with-2xt4 to find out how you can finetune your models using BOTH of the Kaggle provided GPUs. |