yujingfeng commited on
Commit
a9bdb2b
·
verified ·
1 Parent(s): b97863a

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +38 -94
config.json CHANGED
@@ -1,105 +1,49 @@
1
  {
 
2
  "architectures": [
3
- "Qwen2_5_VLForConditionalGeneration"
4
  ],
5
- "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
- "eos_token_id": 151645,
8
- "hidden_act": "silu",
9
- "hidden_size": 3584,
10
- "image_token_id": 151655,
 
 
 
 
11
  "initializer_range": 0.02,
12
- "intermediate_size": 18944,
13
- "max_position_embeddings": 128000,
14
- "max_window_layers": 28,
 
15
  "model_type": "qwen2_5_vl",
16
- "num_attention_heads": 28,
17
- "num_hidden_layers": 28,
18
- "num_key_value_heads": 4,
19
- "rms_norm_eps": 1e-06,
20
- "rope_scaling": {
21
- "mrope_section": [
22
- 16,
23
- 24,
24
- 24
25
- ],
26
- "rope_type": "default",
27
- "type": "default"
28
- },
29
- "rope_theta": 1000000.0,
30
- "sliding_window": 32768,
31
- "text_config": {
32
- "architectures": [
33
- "Qwen2_5_VLForConditionalGeneration"
34
- ],
35
- "attention_dropout": 0.0,
36
- "bos_token_id": 151643,
37
- "eos_token_id": 151645,
38
- "hidden_act": "silu",
39
- "hidden_size": 3584,
40
- "image_token_id": null,
41
- "initializer_range": 0.02,
42
- "intermediate_size": 18944,
43
- "max_position_embeddings": 128000,
44
- "max_window_layers": 28,
45
- "model_type": "qwen2_5_vl_text",
46
- "num_attention_heads": 28,
47
- "num_hidden_layers": 28,
48
- "num_key_value_heads": 4,
49
- "rms_norm_eps": 1e-06,
50
- "rope_scaling": {
51
- "mrope_section": [
52
- 16,
53
- 24,
54
- 24
55
- ],
56
- "rope_type": "default",
57
- "type": "default"
58
- },
59
- "rope_theta": 1000000.0,
60
- "sliding_window": 32768,
61
- "torch_dtype": "bfloat16",
62
- "use_cache": true,
63
- "use_sliding_window": false,
64
- "video_token_id": null,
65
- "vision_end_token_id": 151653,
66
- "vision_start_token_id": 151652,
67
- "vision_token_id": 151654,
68
- "vocab_size": 152064
69
- },
70
  "tie_word_embeddings": false,
 
71
  "torch_dtype": "bfloat16",
72
- "transformers_version": "4.52.4",
73
  "use_cache": true,
74
- "use_sliding_window": false,
75
- "video_token_id": 151656,
76
- "vision_config": {
77
- "depth": 32,
78
- "fullatt_block_indexes": [
79
- 7,
80
- 15,
81
- 23,
82
- 31
83
- ],
84
- "hidden_act": "silu",
85
- "hidden_size": 1280,
86
- "in_channels": 3,
87
- "in_chans": 3,
88
- "initializer_range": 0.02,
89
- "intermediate_size": 3420,
90
- "model_type": "qwen2_5_vl",
91
- "num_heads": 16,
92
- "out_hidden_size": 3584,
93
  "patch_size": 14,
94
- "spatial_merge_size": 2,
95
- "spatial_patch_size": 14,
96
- "temporal_patch_size": 2,
97
- "tokens_per_second": 2,
98
- "torch_dtype": "bfloat16",
99
- "window_size": 112
100
  },
101
- "vision_end_token_id": 151653,
102
- "vision_start_token_id": 151652,
103
- "vision_token_id": 151654,
104
- "vocab_size": 152064
105
  }
 
1
  {
2
+ "_name_or_path": "./",
3
  "architectures": [
4
+ "QWenVLChatModel"
5
  ],
6
+ "attn_dropout_prob": 0.0,
7
+ "auto_map": {
8
+ "AutoConfig": "configuration_qwen.QWenConfig",
9
+ "AutoModelForCausalLM": "modeling_qwen.QWenLMHeadModel"
10
+ },
11
+ "bf16": false,
12
+ "emb_dropout_prob": 0.0,
13
+ "fp16": false,
14
+ "fp32": false,
15
+ "hidden_size": 4096,
16
  "initializer_range": 0.02,
17
+ "intermediate_size": 22016,
18
+ "kv_channels": 128,
19
+ "layer_norm_epsilon": 1e-06,
20
+ "max_position_embeddings": 8192,
21
  "model_type": "qwen2_5_vl",
22
+ "no_bias": true,
23
+ "num_attention_heads": 32,
24
+ "num_hidden_layers": 32,
25
+ "onnx_safe": null,
26
+ "rotary_emb_base": 10000,
27
+ "rotary_pct": 1.0,
28
+ "scale_attn_weights": true,
29
+ "seq_length": 2048,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
30
  "tie_word_embeddings": false,
31
+ "tokenizer_type": "QWenTokenizer",
32
  "torch_dtype": "bfloat16",
33
+ "transformers_version": "4.31.0",
34
  "use_cache": true,
35
+ "use_dynamic_ntk": true,
36
+ "use_flash_attn": false,
37
+ "use_logn_attn": true,
38
+ "visual": {
39
+ "heads": 16,
40
+ "image_size": 448,
41
+ "image_start_id": 151857,
42
+ "layers": 48,
43
+ "mlp_ratio": 4.9231,
44
+ "output_dim": 4096,
 
 
 
 
 
 
 
 
 
45
  "patch_size": 14,
46
+ "width": 1664
 
 
 
 
 
47
  },
48
+ "vocab_size": 151936
 
 
 
49
  }