| { | |
| "adapter_path": "adapters", | |
| "lora_layers": 8, | |
| "lora_parameters": { | |
| "rank": 16, | |
| "alpha": 16, | |
| "dropout": 0.0, | |
| "scale": 1.0 | |
| } | |
| } |
| { | |
| "adapter_path": "adapters", | |
| "lora_layers": 8, | |
| "lora_parameters": { | |
| "rank": 16, | |
| "alpha": 16, | |
| "dropout": 0.0, | |
| "scale": 1.0 | |
| } | |
| } |