| { | |
| "adapters": { | |
| "0": "msfm/llm-jp-3-1.8b_adapter-1", | |
| "1": "msfm/llm-jp-3-1.8b_adapter-2" | |
| }, | |
| "auto_mapping": null, | |
| "base_model_name_or_path": "llm-jp/llm-jp-3-1.8b", | |
| "enable_softmax": true, | |
| "enable_softmax_topk": false, | |
| "global_scaling_weight": 1.0, | |
| "hidden_size": 2048, | |
| "inference_mode": true, | |
| "layerwise_scalings": false, | |
| "peft_type": "XLORA", | |
| "revision": null, | |
| "scaling_pass_value": 0.0, | |
| "softmax_temperature": 1.0, | |
| "task_type": "CAUSAL_LM", | |
| "top_k_lora": null, | |
| "use_trainable_adapters": false, | |
| "xlora_depth": 4, | |
| "xlora_dropout_p": 0.2, | |
| "xlora_size": 2048 | |
| } |