Spaces:
Running
Running
Add Qwen2.5 1.5B as default extraction model
Browse files- Added qwen2.5_1.5b: 1.5B params, 128K context, non-reasoning
- Set as DEFAULT_EXTRACTION_MODEL (replaced granite_350m)
- Qwen2.5 has better instruction following for structured extraction
- Temperature 0.2 for focused but willing extraction
Granite models removed from default due to extraction failures:
- granite_350m: Only extracted from Window 1
- granite_4_0_1b_q4: Returned empty JSON for Windows 2-6
app.py
CHANGED
|
@@ -801,6 +801,22 @@ EXTRACTION_MODELS = {
|
|
| 801 |
"repeat_penalty": 1.0,
|
| 802 |
},
|
| 803 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 804 |
"granite_4_0_1b_q4": {
|
| 805 |
"name": "Granite-4.0 1B (32K Context)",
|
| 806 |
"repo_id": "unsloth/granite-4.0-h-1b-GGUF",
|
|
@@ -863,7 +879,7 @@ EXTRACTION_MODELS = {
|
|
| 863 |
},
|
| 864 |
}
|
| 865 |
|
| 866 |
-
DEFAULT_EXTRACTION_MODEL = "
|
| 867 |
|
| 868 |
|
| 869 |
# ===== ADVANCED MODE: SYNTHESIS MODELS REGISTRY (16 models, 1B-30B) =====
|
|
|
|
| 801 |
"repeat_penalty": 1.0,
|
| 802 |
},
|
| 803 |
},
|
| 804 |
+
"qwen2.5_1.5b": {
|
| 805 |
+
"name": "Qwen2.5 1.5B (128K Context)",
|
| 806 |
+
"repo_id": "Qwen/Qwen2.5-1.5B-Instruct-GGUF",
|
| 807 |
+
"filename": "*Q4_K_M.gguf",
|
| 808 |
+
"max_context": 131072,
|
| 809 |
+
"default_n_ctx": 4096,
|
| 810 |
+
"params_size": "1.5B",
|
| 811 |
+
"supports_reasoning": False,
|
| 812 |
+
"supports_toggle": False,
|
| 813 |
+
"inference_settings": {
|
| 814 |
+
"temperature": 0.2,
|
| 815 |
+
"top_p": 0.9,
|
| 816 |
+
"top_k": 30,
|
| 817 |
+
"repeat_penalty": 1.0,
|
| 818 |
+
},
|
| 819 |
+
},
|
| 820 |
"granite_4_0_1b_q4": {
|
| 821 |
"name": "Granite-4.0 1B (32K Context)",
|
| 822 |
"repo_id": "unsloth/granite-4.0-h-1b-GGUF",
|
|
|
|
| 879 |
},
|
| 880 |
}
|
| 881 |
|
| 882 |
+
DEFAULT_EXTRACTION_MODEL = "qwen2.5_1.5b"
|
| 883 |
|
| 884 |
|
| 885 |
# ===== ADVANCED MODE: SYNTHESIS MODELS REGISTRY (16 models, 1B-30B) =====
|