Upload convert_vision_model.py with huggingface_hub
Browse files- convert_vision_model.py +25 -7
convert_vision_model.py
CHANGED
|
@@ -109,13 +109,31 @@ if IS_LORA:
|
|
| 109 |
with open(adapter_config_path) as f:
|
| 110 |
adapter_config = json.load(f)
|
| 111 |
|
| 112 |
-
# Load base model with specific class
|
| 113 |
-
model_classes = [
|
| 114 |
-
|
| 115 |
-
|
| 116 |
-
|
| 117 |
-
|
| 118 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 119 |
|
| 120 |
base_model = None
|
| 121 |
for class_name, module in model_classes:
|
|
|
|
| 109 |
with open(adapter_config_path) as f:
|
| 110 |
adapter_config = json.load(f)
|
| 111 |
|
| 112 |
+
# Load base model with specific class - detect from model name
|
| 113 |
+
model_classes = []
|
| 114 |
+
base_lower = BASE_MODEL.lower()
|
| 115 |
+
if "ministral" in base_lower or "mistral" in base_lower:
|
| 116 |
+
model_classes = [
|
| 117 |
+
("Mistral3ForConditionalGeneration", "transformers"),
|
| 118 |
+
("AutoModelForCausalLM", "transformers"),
|
| 119 |
+
]
|
| 120 |
+
elif "glm" in base_lower:
|
| 121 |
+
model_classes = [
|
| 122 |
+
("Glm4vForConditionalGeneration", "transformers"),
|
| 123 |
+
("AutoModelForVision2Seq", "transformers"),
|
| 124 |
+
]
|
| 125 |
+
elif "gemma" in base_lower:
|
| 126 |
+
model_classes = [
|
| 127 |
+
("Gemma3ForConditionalGeneration", "transformers"),
|
| 128 |
+
("AutoModelForVision2Seq", "transformers"),
|
| 129 |
+
]
|
| 130 |
+
else:
|
| 131 |
+
model_classes = [
|
| 132 |
+
("AutoModelForCausalLM", "transformers"),
|
| 133 |
+
("AutoModelForVision2Seq", "transformers"),
|
| 134 |
+
]
|
| 135 |
+
|
| 136 |
+
print(f" Detected model type, trying: {[c[0] for c in model_classes]}")
|
| 137 |
|
| 138 |
base_model = None
|
| 139 |
for class_name, module in model_classes:
|