| license: apache-2.0 | |
| language: | |
| - zh | |
| - en | |
| base_model: | |
| - Qwen/Qwen2-7B-Instruct | |
| - meta-llama/Llama-3.1-8B-Instruct | |
| pipeline_tag: text-generation | |
| ## Training procedure | |
| - total_batch_size: 32 | |
| - epoch: 3 | |
| - lr: 1.0e-4 | |
| - warm-up rate: 0.1 | |
| - type: Lora | |
| ## Framework versions | |
| - LLaMA-Factory: v0.9.0 | |
| ## Paper | |
| - link: arxiv.org/abs/2412.04905 | |
| ## Data | |
| - link: https://github.com/MozerWang/DEMO | |