Delete config.json

#1
by XshinzoX - opened
Files changed (1) hide show
  1. config.json +0 -131
config.json DELETED
@@ -1,131 +0,0 @@
1
- {
2
- "_name_or_path": "unsloth/Qwen2-VL-2B-Instruct",
3
- "architectures": [
4
- "Qwen2VLForConditionalGeneration"
5
- ],
6
- "attention_dropout": 0.0,
7
- "eos_token_id": 151645,
8
- "hidden_act": "silu",
9
- "hidden_size": 1536,
10
- "image_token_id": 151655,
11
- "initializer_range": 0.02,
12
- "intermediate_size": 8960,
13
- "max_position_embeddings": 32768,
14
- "max_window_layers": 28,
15
- "model_type": "qwen2_vl",
16
- "num_attention_heads": 12,
17
- "num_hidden_layers": 28,
18
- "num_key_value_heads": 2,
19
- "pad_token_id": 151654,
20
- "quantization_config": {
21
- "_load_in_4bit": true,
22
- "_load_in_8bit": false,
23
- "bnb_4bit_compute_dtype": "bfloat16",
24
- "bnb_4bit_quant_storage": "uint8",
25
- "bnb_4bit_quant_type": "nf4",
26
- "bnb_4bit_use_double_quant": true,
27
- "llm_int8_enable_fp32_cpu_offload": false,
28
- "llm_int8_has_fp16_weight": false,
29
- "llm_int8_skip_modules": [
30
- "lm_head",
31
- "multi_modal_projector",
32
- "merger",
33
- "modality_projection",
34
- "model.layers.1.mlp",
35
- "visual.blocks.29.mlp",
36
- "visual.blocks.21.mlp",
37
- "visual.blocks.24.attn",
38
- "visual.blocks.26.mlp",
39
- "visual.blocks.23.attn",
40
- "visual.blocks.18.mlp",
41
- "visual.blocks.19.mlp",
42
- "visual.blocks.22.attn",
43
- "visual.blocks.17.mlp",
44
- "visual.blocks.25.attn",
45
- "visual.blocks.24.mlp",
46
- "visual.blocks.26.attn",
47
- "visual.blocks.29.attn",
48
- "visual.blocks.21.attn",
49
- "visual.blocks.22.mlp",
50
- "visual.blocks.14.mlp",
51
- "visual.blocks.23.mlp",
52
- "visual.blocks.14.attn",
53
- "visual.merger.mlp",
54
- "visual.blocks.16.mlp",
55
- "visual.blocks.20.mlp",
56
- "visual.blocks.25.mlp",
57
- "visual.blocks.13.mlp",
58
- "visual.blocks.18.attn",
59
- "visual.blocks.20.attn",
60
- "visual.blocks.17.attn",
61
- "visual.blocks.15.mlp",
62
- "visual.blocks.27.attn",
63
- "visual.blocks.28.attn",
64
- "visual.blocks.12.mlp",
65
- "visual.blocks.16.attn",
66
- "visual.blocks.19.attn",
67
- "visual.blocks.15.attn",
68
- "visual.blocks.11.attn",
69
- "visual.blocks.13.attn",
70
- "visual.blocks.12.attn",
71
- "visual.blocks.10.mlp",
72
- "visual.blocks.11.mlp",
73
- "visual.blocks.10.attn",
74
- "visual.blocks.9.mlp",
75
- "visual.blocks.5.attn",
76
- "visual.blocks.9.attn",
77
- "visual.blocks.4.attn",
78
- "visual.blocks.6.mlp",
79
- "visual.blocks.7.mlp",
80
- "visual.blocks.1.attn",
81
- "visual.blocks.6.attn",
82
- "visual.blocks.7.attn",
83
- "visual.blocks.5.mlp",
84
- "visual.blocks.8.attn",
85
- "visual.blocks.8.mlp",
86
- "visual.blocks.4.mlp",
87
- "visual.blocks.3.attn",
88
- "visual.blocks.3.mlp",
89
- "visual.blocks.1.mlp",
90
- "visual.blocks.2.mlp",
91
- "visual.blocks.2.attn",
92
- "visual.blocks.0.attn",
93
- "visual.blocks.0.mlp",
94
- "visual.blocks.31.mlp"
95
- ],
96
- "llm_int8_threshold": 6.0,
97
- "load_in_4bit": true,
98
- "load_in_8bit": false,
99
- "quant_method": "bitsandbytes"
100
- },
101
- "rms_norm_eps": 1e-06,
102
- "rope_scaling": {
103
- "mrope_section": [
104
- 16,
105
- 24,
106
- 24
107
- ],
108
- "rope_type": "default",
109
- "type": "default"
110
- },
111
- "rope_theta": 1000000.0,
112
- "sliding_window": 32768,
113
- "tie_word_embeddings": true,
114
- "torch_dtype": "bfloat16",
115
- "transformers_version": "4.49.0",
116
- "unsloth_fixed": true,
117
- "use_cache": true,
118
- "use_sliding_window": false,
119
- "video_token_id": 151656,
120
- "vision_config": {
121
- "hidden_size": 1536,
122
- "in_chans": 3,
123
- "model_type": "qwen2_vl",
124
- "spatial_patch_size": 14,
125
- "torch_dtype": "bfloat16"
126
- },
127
- "vision_end_token_id": 151653,
128
- "vision_start_token_id": 151652,
129
- "vision_token_id": 151654,
130
- "vocab_size": 151936
131
- }