Update README.md
Browse files
README.md
CHANGED
@@ -19,8 +19,12 @@ tags:
|
|
19 |
|
20 |
Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these two datasets:
|
21 |
[ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full)
|
22 |
-
[Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (
|
23 |
|
|
|
|
|
|
|
|
|
24 |
15062 rows in total with a 10% validation split.
|
25 |
|
26 |
Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template.
|
@@ -42,7 +46,7 @@ This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unsloth
|
|
42 |
|
43 |
# Training details
|
44 |
|
45 |
-
About 1
|
46 |
|
47 |
Rank: 128
|
48 |
|
@@ -74,15 +78,23 @@ TrainingArguments(
|
|
74 |
|
75 |
|
76 |
Step Training Loss Validation Loss
|
77 |
-
|
78 |
-
20
|
79 |
-
|
80 |
-
40
|
81 |
-
|
82 |
-
60 0.
|
83 |
-
|
84 |
-
80 0.
|
85 |
-
|
86 |
-
100 0.
|
87 |
-
|
88 |
-
120 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
19 |
|
20 |
Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these two datasets:
|
21 |
[ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full)
|
22 |
+
[Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (10%)
|
23 |
|
24 |
+
preview 2:
|
25 |
+
26210 rows, of which ca 1000 are from my own multi response dataset
|
26 |
+
|
27 |
+
preview 1:
|
28 |
15062 rows in total with a 10% validation split.
|
29 |
|
30 |
Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template.
|
|
|
46 |
|
47 |
# Training details
|
48 |
|
49 |
+
About 1.5 epochs. It's probably a bit overfitting and I should introduce some general coding questions to my validation set to ensure it doesn't lose too much general performance.
|
50 |
|
51 |
Rank: 128
|
52 |
|
|
|
78 |
|
79 |
|
80 |
Step Training Loss Validation Loss
|
81 |
+
|
82 |
+
20 2.043000 1.197104
|
83 |
+
|
84 |
+
40 1.087300 0.933553
|
85 |
+
|
86 |
+
60 0.942200 0.890801
|
87 |
+
|
88 |
+
80 0.865600 0.866198
|
89 |
+
|
90 |
+
100 0.851400 0.849733
|
91 |
+
|
92 |
+
120 0.812900 0.837039
|
93 |
+
|
94 |
+
140 0.812400 0.827064
|
95 |
+
|
96 |
+
160 0.817300 0.818410
|
97 |
+
|
98 |
+
180 0.802600 0.810163
|
99 |
+
|
100 |
+
200 0.788600 0.803399
|