neph1 commited on
Commit
6427b16
·
verified ·
1 Parent(s): e8f46b4

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +26 -14
README.md CHANGED
@@ -19,8 +19,12 @@ tags:
19
 
20
  Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these two datasets:
21
  [ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full)
22
- [Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (5%)
23
 
 
 
 
 
24
  15062 rows in total with a 10% validation split.
25
 
26
  Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template.
@@ -42,7 +46,7 @@ This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unsloth
42
 
43
  # Training details
44
 
45
- About 1 epoch.
46
 
47
  Rank: 128
48
 
@@ -74,15 +78,23 @@ TrainingArguments(
74
 
75
 
76
  Step Training Loss Validation Loss
77
- 10 2.097300 1.165832
78
- 20 1.058100 1.013441
79
- 30 0.898500 0.969640
80
- 40 0.866600 0.943687
81
- 50 0.847300 0.926879
82
- 60 0.838200 0.903914
83
- 70 0.797600 0.888580
84
- 80 0.777700 0.873389
85
- 90 0.793900 0.859501
86
- 100 0.725500 0.846339
87
- 110 0.739400 0.843786
88
- 120 0.675200 0.833775
 
 
 
 
 
 
 
 
 
19
 
20
  Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these two datasets:
21
  [ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full)
22
+ [Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (10%)
23
 
24
+ preview 2:
25
+ 26210 rows, of which ca 1000 are from my own multi response dataset
26
+
27
+ preview 1:
28
  15062 rows in total with a 10% validation split.
29
 
30
  Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template.
 
46
 
47
  # Training details
48
 
49
+ About 1.5 epochs. It's probably a bit overfitting and I should introduce some general coding questions to my validation set to ensure it doesn't lose too much general performance.
50
 
51
  Rank: 128
52
 
 
78
 
79
 
80
  Step Training Loss Validation Loss
81
+
82
+ 20 2.043000 1.197104
83
+
84
+ 40 1.087300 0.933553
85
+
86
+ 60 0.942200 0.890801
87
+
88
+ 80 0.865600 0.866198
89
+
90
+ 100 0.851400 0.849733
91
+
92
+ 120 0.812900 0.837039
93
+
94
+ 140 0.812400 0.827064
95
+
96
+ 160 0.817300 0.818410
97
+
98
+ 180 0.802600 0.810163
99
+
100
+ 200 0.788600 0.803399