Upload folder using huggingface_hub
Browse files- README.md +1 -1
- config.json +1 -1
- generation_config.json +1 -1
- model.safetensors +1 -1
README.md
CHANGED
|
@@ -17,4 +17,4 @@ Strict JSON structured output specialist. Produces valid JSON function calls.
|
|
| 17 |
## Training
|
| 18 |
- Method: LoRA r=16, alpha=32 (merged)
|
| 19 |
- Samples: 15,000
|
| 20 |
-
- Time:
|
|
|
|
| 17 |
## Training
|
| 18 |
- Method: LoRA r=16, alpha=32 (merged)
|
| 19 |
- Samples: 15,000
|
| 20 |
+
- Time: 1.7 min
|
config.json
CHANGED
|
@@ -15,7 +15,7 @@
|
|
| 15 |
"router_aux_loss_coef": 0.01,
|
| 16 |
"shared_expert_intermediate_size": 1408,
|
| 17 |
"top_k": 2,
|
| 18 |
-
"transformers_version": "5.
|
| 19 |
"use_cache": false,
|
| 20 |
"use_checkpointing": false,
|
| 21 |
"vocab_size": 50304
|
|
|
|
| 15 |
"router_aux_loss_coef": 0.01,
|
| 16 |
"shared_expert_intermediate_size": 1408,
|
| 17 |
"top_k": 2,
|
| 18 |
+
"transformers_version": "5.1.0",
|
| 19 |
"use_cache": false,
|
| 20 |
"use_checkpointing": false,
|
| 21 |
"vocab_size": 50304
|
generation_config.json
CHANGED
|
@@ -2,5 +2,5 @@
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"output_attentions": false,
|
| 4 |
"output_hidden_states": false,
|
| 5 |
-
"transformers_version": "5.
|
| 6 |
}
|
|
|
|
| 2 |
"_from_model_config": true,
|
| 3 |
"output_attentions": false,
|
| 4 |
"output_hidden_states": false,
|
| 5 |
+
"transformers_version": "5.1.0"
|
| 6 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 543985840
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b3445b2682fce85d45f22fa5120af0626268d889c6203f6d1e804df5e0d1a694
|
| 3 |
size 543985840
|