Update README.md
Browse files
README.md
CHANGED
|
@@ -19,6 +19,7 @@ tags:
|
|
| 19 |
|
| 20 |
This repository provides a **LoRA adapter** fine-tuned from
|
| 21 |
**Matukaze/test105** using **LoRA + Unsloth**.
|
|
|
|
| 22 |
|
| 23 |
This repository contains **LoRA adapter weights only**.
|
| 24 |
The base model must be loaded separately.
|
|
@@ -40,6 +41,11 @@ tool use, and recovery from errors.
|
|
| 40 |
- Epochs: 0
|
| 41 |
- Learning rate: 5e-06
|
| 42 |
- LoRA: r=32, alpha=128
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 43 |
|
| 44 |
## Usage
|
| 45 |
|
|
|
|
| 19 |
|
| 20 |
This repository provides a **LoRA adapter** fine-tuned from
|
| 21 |
**Matukaze/test105** using **LoRA + Unsloth**.
|
| 22 |
+
This model initialised from Qwen2.5-7B-Instruct.
|
| 23 |
|
| 24 |
This repository contains **LoRA adapter weights only**.
|
| 25 |
The base model must be loaded separately.
|
|
|
|
| 41 |
- Epochs: 0
|
| 42 |
- Learning rate: 5e-06
|
| 43 |
- LoRA: r=32, alpha=128
|
| 44 |
+
- This model was fine-tuned sequentially using LoRA
|
| 45 |
+
- First stage:fine-tuned on DBBench SFT dataset and merged.
|
| 46 |
+
- Second stage:furter fine-tuned on ALFWorld SFT datasset and merged.
|
| 47 |
+
|
| 48 |
+
-This model architecture remains unchanged.
|
| 49 |
|
| 50 |
## Usage
|
| 51 |
|