| { | |
| "config": { | |
| "architecture": "prefix_tuning", | |
| "bottleneck_size": 512, | |
| "cross_prefix": true, | |
| "dropout": 0.0, | |
| "encoder_prefix": true, | |
| "flat": false, | |
| "leave_out": [], | |
| "non_linearity": "tanh", | |
| "prefix_length": 30, | |
| "shared_gating": true, | |
| "use_gating": false | |
| }, | |
| "config_id": "648bf22f5afeaaa6", | |
| "hidden_size": 2048, | |
| "model_class": "LlamaForCausalLM", | |
| "model_name": "meta-llama/Llama-3.2-1B-Instruct", | |
| "model_type": "llama", | |
| "name": "llama23-1b-prefix-is", | |
| "version": "adapters.1.0.0" | |
| } |