Update README.md
Browse files
README.md
CHANGED
@@ -16,8 +16,8 @@ This model is a fine-tuned version of [microsoft/phi-2](https://huggingface.co/m
|
|
16 |
I think SPIN not only can use on a SFT model, but also it can use on a pretrained model.
|
17 |
Therefore, I use SPIN on a pretrained model microsoft/phi-2. And I get a higher score better than origin pretrained model. You can check the [open llm leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
|
18 |
|
19 |
-
|
20 |
-
pretrain -> dpo(spin) -> sft -> dpo(spin)
|
21 |
|
22 |
## Training procedure
|
23 |
|
|
|
16 |
I think SPIN not only can use on a SFT model, but also it can use on a pretrained model.
|
17 |
Therefore, I use SPIN on a pretrained model microsoft/phi-2. And I get a higher score better than origin pretrained model. You can check the [open llm leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard).
|
18 |
|
19 |
+
**I Think the best paradigm for training a conversational Large Language Model (LLM):
|
20 |
+
pretrain -> dpo(spin) -> sft -> dpo(spin)**
|
21 |
|
22 |
## Training procedure
|
23 |
|