Update README.md
Browse files
README.md
CHANGED
|
@@ -8,22 +8,6 @@ tags:
|
|
| 8 |
- llama3
|
| 9 |
---
|
| 10 |
|
| 11 |
-
---
|
| 12 |
-
frameworks:
|
| 13 |
-
- Pytorch
|
| 14 |
-
license: Apache License 2.0
|
| 15 |
-
tasks:
|
| 16 |
-
- chatbot
|
| 17 |
-
|
| 18 |
-
language:
|
| 19 |
-
- cn
|
| 20 |
-
|
| 21 |
-
tags:
|
| 22 |
-
- RL-tuned
|
| 23 |
-
|
| 24 |
-
tools:
|
| 25 |
-
- vllm
|
| 26 |
-
---
|
| 27 |
Github:https://github.com/CrazyBoyM/llama3-Chinese-chat
|
| 28 |
放出训练配方细节供网友参考分享:
|
| 29 |
DPO(beta 0.5) + lora rank128, alpha256 + 打开"lm_head", "input_layernorm", "post_attention_layernorm", "norm"层训练。
|
|
|
|
| 8 |
- llama3
|
| 9 |
---
|
| 10 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 11 |
Github:https://github.com/CrazyBoyM/llama3-Chinese-chat
|
| 12 |
放出训练配方细节供网友参考分享:
|
| 13 |
DPO(beta 0.5) + lora rank128, alpha256 + 打开"lm_head", "input_layernorm", "post_attention_layernorm", "norm"层训练。
|