| { | |
| "model_type": "llama", | |
| "architectures": [ | |
| "FastLanguageModel" | |
| ], | |
| "max_seq_length": 2048, | |
| "hidden_size": 4096, | |
| "num_attention_heads": 32, | |
| "num_hidden_layers": 32, | |
| "r": 16, | |
| "lora_alpha": 16, | |
| "lora_dropout": 0, | |
| "use_gradient_checkpointing": true | |
| } |
| { | |
| "model_type": "llama", | |
| "architectures": [ | |
| "FastLanguageModel" | |
| ], | |
| "max_seq_length": 2048, | |
| "hidden_size": 4096, | |
| "num_attention_heads": 32, | |
| "num_hidden_layers": 32, | |
| "r": 16, | |
| "lora_alpha": 16, | |
| "lora_dropout": 0, | |
| "use_gradient_checkpointing": true | |
| } |