Update README.md
Browse files
README.md
CHANGED
|
@@ -7,8 +7,15 @@ datasets:
|
|
| 7 |
This repo reproduced [tloen/alpaca-lora-7b](https://huggingface.co/tloen/alpaca-lora-7b)
|
| 8 |
fit on the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset.
|
| 9 |
|
| 10 |
-
|
| 11 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 12 |
|
| 13 |
This version of the weights was trained with the following hyperparameters:
|
| 14 |
|
|
|
|
| 7 |
This repo reproduced [tloen/alpaca-lora-7b](https://huggingface.co/tloen/alpaca-lora-7b)
|
| 8 |
fit on the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset.
|
| 9 |
|
| 10 |
+
4x H100 training for about 1h15min, details in [W&B link](https://wandb.ai/sharpbai/alpaca-lora-reproduce/runs/08ulvstd), there is a hyperparameter of val_set_size=500
|
| 11 |
+
|
| 12 |
+
4 x 4090 training for about 4h35min, details in [W&B link](https://wandb.ai/sharpbai/alpaca-lora-reproduce/runs/ws16av1u), all key hyperparameters are the same
|
| 13 |
+
|
| 14 |
+
To optimize the running speed, I change these code
|
| 15 |
+
|
| 16 |
+
- `load_in_8bits=False` to use 16bit finetune
|
| 17 |
+
- comment `model = prepare_model_for_int8_training` to not turn some parameters to fp32 and turn off gradient checkpointing
|
| 18 |
+
- for 4090 enable gradient checkpointing, add `model.gradient_checkpointing_enable()` and `model.enable_input_require_grads()`
|
| 19 |
|
| 20 |
This version of the weights was trained with the following hyperparameters:
|
| 21 |
|