Update README.md
Browse files
README.md
CHANGED
|
@@ -19,8 +19,12 @@ tags:
|
|
| 19 |
|
| 20 |
Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these two datasets:
|
| 21 |
[ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full)
|
| 22 |
-
[Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (
|
| 23 |
|
|
|
|
|
|
|
|
|
|
|
|
|
| 24 |
15062 rows in total with a 10% validation split.
|
| 25 |
|
| 26 |
Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template.
|
|
@@ -42,7 +46,7 @@ This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unsloth
|
|
| 42 |
|
| 43 |
# Training details
|
| 44 |
|
| 45 |
-
About 1
|
| 46 |
|
| 47 |
Rank: 128
|
| 48 |
|
|
@@ -74,15 +78,23 @@ TrainingArguments(
|
|
| 74 |
|
| 75 |
|
| 76 |
Step Training Loss Validation Loss
|
| 77 |
-
|
| 78 |
-
20
|
| 79 |
-
|
| 80 |
-
40
|
| 81 |
-
|
| 82 |
-
60 0.
|
| 83 |
-
|
| 84 |
-
80 0.
|
| 85 |
-
|
| 86 |
-
100 0.
|
| 87 |
-
|
| 88 |
-
120 0.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 19 |
|
| 20 |
Qwen2.5-Coder-7B-Instruct trained on a merged dataset of Unity3d q&a from these two datasets:
|
| 21 |
[ibranze/codellama_unity3d_v2](https://huggingface.co/datasets/ibranze/codellama_unity3d_v2) (Full)
|
| 22 |
+
[Hypersniper/unity_api_2022_3](https://huggingface.co/datasets/Hypersniper/unity_api_2022_3) (10%)
|
| 23 |
|
| 24 |
+
preview 2:
|
| 25 |
+
26210 rows, of which ca 1000 are from my own multi response dataset
|
| 26 |
+
|
| 27 |
+
preview 1:
|
| 28 |
15062 rows in total with a 10% validation split.
|
| 29 |
|
| 30 |
Trained with native chat template (minus tools usage, see this issue: https://github.com/unslothai/unsloth/issues/1053). With a little superficial testing done, it seems to respond well to the mistral template.
|
|
|
|
| 46 |
|
| 47 |
# Training details
|
| 48 |
|
| 49 |
+
About 1.5 epochs. It's probably a bit overfitting and I should introduce some general coding questions to my validation set to ensure it doesn't lose too much general performance.
|
| 50 |
|
| 51 |
Rank: 128
|
| 52 |
|
|
|
|
| 78 |
|
| 79 |
|
| 80 |
Step Training Loss Validation Loss
|
| 81 |
+
|
| 82 |
+
20 2.043000 1.197104
|
| 83 |
+
|
| 84 |
+
40 1.087300 0.933553
|
| 85 |
+
|
| 86 |
+
60 0.942200 0.890801
|
| 87 |
+
|
| 88 |
+
80 0.865600 0.866198
|
| 89 |
+
|
| 90 |
+
100 0.851400 0.849733
|
| 91 |
+
|
| 92 |
+
120 0.812900 0.837039
|
| 93 |
+
|
| 94 |
+
140 0.812400 0.827064
|
| 95 |
+
|
| 96 |
+
160 0.817300 0.818410
|
| 97 |
+
|
| 98 |
+
180 0.802600 0.810163
|
| 99 |
+
|
| 100 |
+
200 0.788600 0.803399
|