Update README.md
Browse files
README.md
CHANGED
|
@@ -72,4 +72,11 @@ E. this situation can and will be changed
|
|
| 72 |
F. let us not wallow in the valley of despair
|
| 73 |
```
|
| 74 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 75 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|
|
|
|
| 72 |
F. let us not wallow in the valley of despair
|
| 73 |
```
|
| 74 |
|
| 75 |
+
### Lessons Learned
|
| 76 |
+
|
| 77 |
+
- When adding new tokens to the model, LoRA will be much worse. Use full fine-tuning to get better results.
|
| 78 |
+
- Be very careful about chat templates. Every character/new line/space matters and not following that can make the model have worse performance.
|
| 79 |
+
- For Qwen base models, leave the `<|endoftext|>` as the EOS token. Then you can train it to use other tokens like `<|im_end|>`. If you set the EOS token to `<|im_end|>`, the model will get confused.
|
| 80 |
+
- For Qwen models in general, always put the `<|endoftext|>` at the end of each training example.
|
| 81 |
+
|
| 82 |
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
|