|
|
--- |
|
|
license: bigscience-openrail-m |
|
|
datasets: |
|
|
- iamplus/Instruction_Tuning |
|
|
--- |
|
|
First Version of Instruction Tuned Bloomz-7B1 model on ChatGPT dataset (85k data) using ***HF Deepspeed*** |
|
|
|
|
|
**Base Model:** bigscience/bloomz-7b1 |
|
|
|
|
|
**Training Details :** |
|
|
* Epochs: 5 |
|
|
* Batch Size : 5 instantaneous per device x 2 gradient accumulation steps x 8 gpus = 80 |
|
|
* Max Length : 512 |
|
|
* Weight Decay : 0 |
|
|
* Learning Rate : 5e-5 |
|
|
* Learning Rate Scheduler Type : Linear |
|
|
* Number of warmup steps : 0 |
|
|
* Machine : 8xA100 80GB |
|
|
|
|
|
**Dataset Details :** |
|
|
|
|
|
Dataset : iamplus/Instruction_Tuning |
|
|
|
|
|
Files : |
|
|
* chat_gpt_v1.csv |