---
base_model: meta-llama/Llama-3.2-3B
library_name: transformers
model_name: notHumpback-M1
tags:
- generated_from_trainer
- trl
- sft
license: apache-2.0
datasets:
- OpenAssistant/oasst1
- allenai/c4
---

# notHumpback-M1

This model follows the Humpback architecture, proposed in the paper [Self-Alignment with Instruction Backtranslation](https://arxiv.org/pdf/2308.06259) 
by Li et al.

It represents the resulting model after the first iteration of self-curation, which is trained on a small amount of gold data 
and a set of generated data curated by the ["seed model"](https://huggingface.co/Alepach/notHumpback-M0). 

This model can be used for instruction-following.
It may also be used to, again, score the instruction-response pairs 
generated by the ["backward model"](https://huggingface.co/Alepach/notHumpback-Myx) for a second iteration of self-curation. 

Humpback uses instruction backtranslation on a web corpus to generate input-output pairs (self-augmentation), 
creating a richer dataset for fine-tuning models without the need for additional manual annotation.
The model then iteratively curates the created dataset, scoring the pairs by quality, and is then finetuned on the resulting subset 
of all pairs with the highest possible score (self-curation).

Varying from the original paper, this model is a fine-tuned version of [meta-llama/Llama-3.2-3B](https://huggingface.co/meta-llama/Llama-3.2-3B).
It has been trained using [TRL](https://github.com/huggingface/trl).

The dataset used to train this model is a combination of data sampled from the [oasst1](https://huggingface.co/datasets/OpenAssistant/oasst1) 
dataset and the synthetic dataset which was mentioned above. The latter has been created by applying self-augmentation and self-curation 
on 502k entries from the english subset ("en") of the [c4](https://huggingface.co/datasets/allenai/c4) dataset.

For comparison with other methods, the training dataset was limited to 16000 instruction-response pairs.

### Framework versions

- TRL: 0.12.1
- Transformers: 4.46.3
- Pytorch: 2.5.1
- Datasets: 3.1.0
- Tokenizers: 0.20.3

## Citations

Original paper:

```bibtex
@misc{li2023selfalignment,
    title={Self-Alignment with Instruction Backtranslation},
    author={Xian Li and Ping Yu and Chunting Zhou and Timo Schick and Luke Zettlemoyer and Omer Levy and Jason Weston and Mike Lewis},
    year={2023},
    eprint={2308.06259},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}
```

Cite TRL as:
    
```bibtex
@misc{vonwerra2022trl,
	title        = {{TRL: Transformer Reinforcement Learning}},
	author       = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec},
	year         = 2020,
	journal      = {GitHub repository},
	publisher    = {GitHub},
	howpublished = {\url{https://github.com/huggingface/trl}}
}
```