Upload configuration_llava_qwen.py with huggingface_hub
Browse files- configuration_llava_qwen.py +42 -0
configuration_llava_qwen.py
ADDED
|
@@ -0,0 +1,42 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
# Configuration for LLaVA Qwen3 model
|
| 2 |
+
# Upload this file to your HuggingFace repo for trust_remote_code=True
|
| 3 |
+
|
| 4 |
+
from transformers import Qwen2Config
|
| 5 |
+
|
| 6 |
+
try:
|
| 7 |
+
from transformers import Qwen3Config
|
| 8 |
+
QWEN3_AVAILABLE = True
|
| 9 |
+
except ImportError:
|
| 10 |
+
QWEN3_AVAILABLE = False
|
| 11 |
+
Qwen3Config = Qwen2Config # Fallback
|
| 12 |
+
|
| 13 |
+
|
| 14 |
+
class LlavaQwen3Config(Qwen3Config if QWEN3_AVAILABLE else Qwen2Config):
|
| 15 |
+
"""Configuration class for LLaVA Qwen3 model."""
|
| 16 |
+
|
| 17 |
+
model_type = "llava_qwen3"
|
| 18 |
+
|
| 19 |
+
def __init__(
|
| 20 |
+
self,
|
| 21 |
+
mm_vision_tower=None,
|
| 22 |
+
mm_hidden_size=None,
|
| 23 |
+
mm_projector_type="mlp2x_gelu",
|
| 24 |
+
mm_vision_select_layer=-2,
|
| 25 |
+
mm_vision_select_feature="patch",
|
| 26 |
+
mm_patch_merge_type="flat",
|
| 27 |
+
mm_use_im_start_end=False,
|
| 28 |
+
mm_use_im_patch_token=False,
|
| 29 |
+
image_aspect_ratio="pad",
|
| 30 |
+
**kwargs
|
| 31 |
+
):
|
| 32 |
+
super().__init__(**kwargs)
|
| 33 |
+
|
| 34 |
+
self.mm_vision_tower = mm_vision_tower
|
| 35 |
+
self.mm_hidden_size = mm_hidden_size
|
| 36 |
+
self.mm_projector_type = mm_projector_type
|
| 37 |
+
self.mm_vision_select_layer = mm_vision_select_layer
|
| 38 |
+
self.mm_vision_select_feature = mm_vision_select_feature
|
| 39 |
+
self.mm_patch_merge_type = mm_patch_merge_type
|
| 40 |
+
self.mm_use_im_start_end = mm_use_im_start_end
|
| 41 |
+
self.mm_use_im_patch_token = mm_use_im_patch_token
|
| 42 |
+
self.image_aspect_ratio = image_aspect_ratio
|