Luigi commited on
Commit
9f68f0b
·
1 Parent(s): 8b92581

Add Qwen2.5 1.5B as default extraction model

Browse files

- Added qwen2.5_1.5b: 1.5B params, 128K context, non-reasoning
- Set as DEFAULT_EXTRACTION_MODEL (replaced granite_350m)
- Qwen2.5 has better instruction following for structured extraction
- Temperature 0.2 for focused but willing extraction

Granite models removed from default due to extraction failures:
- granite_350m: Only extracted from Window 1
- granite_4_0_1b_q4: Returned empty JSON for Windows 2-6

Files changed (1) hide show
  1. app.py +17 -1
app.py CHANGED
@@ -801,6 +801,22 @@ EXTRACTION_MODELS = {
801
  "repeat_penalty": 1.0,
802
  },
803
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
804
  "granite_4_0_1b_q4": {
805
  "name": "Granite-4.0 1B (32K Context)",
806
  "repo_id": "unsloth/granite-4.0-h-1b-GGUF",
@@ -863,7 +879,7 @@ EXTRACTION_MODELS = {
863
  },
864
  }
865
 
866
- DEFAULT_EXTRACTION_MODEL = "granite_350m"
867
 
868
 
869
  # ===== ADVANCED MODE: SYNTHESIS MODELS REGISTRY (16 models, 1B-30B) =====
 
801
  "repeat_penalty": 1.0,
802
  },
803
  },
804
+ "qwen2.5_1.5b": {
805
+ "name": "Qwen2.5 1.5B (128K Context)",
806
+ "repo_id": "Qwen/Qwen2.5-1.5B-Instruct-GGUF",
807
+ "filename": "*Q4_K_M.gguf",
808
+ "max_context": 131072,
809
+ "default_n_ctx": 4096,
810
+ "params_size": "1.5B",
811
+ "supports_reasoning": False,
812
+ "supports_toggle": False,
813
+ "inference_settings": {
814
+ "temperature": 0.2,
815
+ "top_p": 0.9,
816
+ "top_k": 30,
817
+ "repeat_penalty": 1.0,
818
+ },
819
+ },
820
  "granite_4_0_1b_q4": {
821
  "name": "Granite-4.0 1B (32K Context)",
822
  "repo_id": "unsloth/granite-4.0-h-1b-GGUF",
 
879
  },
880
  }
881
 
882
+ DEFAULT_EXTRACTION_MODEL = "qwen2.5_1.5b"
883
 
884
 
885
  # ===== ADVANCED MODE: SYNTHESIS MODELS REGISTRY (16 models, 1B-30B) =====