--- language: - en - ko license: other license_name: exaone license_link: LICENSE tags: - text-generation-inference - transformers - trl - sft - reasoning - lg-ai - exaone - exaone-3.5 - o1 base_model: LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct datasets: - KingNish/reasoning-base-20k --- # Model Description An uncensored reasoning EXAONE 3.5 model trained on reasoning data. Now with a full epoch! It has been trained using improved training code, and gives an improved performance. Here is what inference code you should use: ```py # DEBUGGING IN PROGRESS, check later ``` - **Trained by:** [Piotr Zalewski](https://huggingface.co/lunahr) - **License:** exaone - **Finetuned from model:** [LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct](https://huggingface.co/LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct) - **Dataset used:** [KingNish/reasoning-base-20k](https://huggingface.co/datasets/KingNish/reasoning-base-20k) This Llama model was trained faster than [Unsloth](https://github.com/unslothai/unsloth) using [custom training code](https://www.kaggle.com/code/piotr25691/distributed-hf-training-with-2xt4). Visit https://www.kaggle.com/code/piotr25691/distributed-hf-training-with-2xt4 to find out how you can finetune your models using BOTH of the Kaggle provided GPUs.