{ "adapters": { "0": "msfm/llm-jp-3-1.8b_adapter-1", "1": "msfm/llm-jp-3-1.8b_adapter-2" }, "auto_mapping": null, "base_model_name_or_path": "llm-jp/llm-jp-3-1.8b", "enable_softmax": true, "enable_softmax_topk": false, "global_scaling_weight": 1.0, "hidden_size": 2048, "inference_mode": true, "layerwise_scalings": false, "peft_type": "XLORA", "revision": null, "scaling_pass_value": 0.0, "softmax_temperature": 1.0, "task_type": "CAUSAL_LM", "top_k_lora": null, "use_trainable_adapters": false, "xlora_depth": 4, "xlora_dropout_p": 0.2, "xlora_size": 2048 }