chnaaam commited on
Commit
c3961cd
·
verified ·
1 Parent(s): 5c8c7ab

Upload config.json

Browse files
Files changed (1) hide show
  1. config.json +19 -19
config.json CHANGED
@@ -4,15 +4,15 @@
4
  ],
5
  "auto_map": {
6
  "AutoConfig": "configuration_ovis.OvisConfig",
7
- "AutoModelForCausalLM": "AIDC-AI/Ovis2-2B--modeling_ovis.Ovis"
8
  },
9
  "conversation_formatter_class": "QwenConversationFormatter",
10
  "disable_tie_weight": false,
11
- "hidden_size": 1536,
12
  "llm_attn_implementation": "flash_attention_2",
13
  "llm_config": {
14
  "_attn_implementation_autoset": true,
15
- "_name_or_path": "Qwen/Qwen2.5-1.5B-Instruct",
16
  "add_cross_attention": false,
17
  "architectures": [
18
  "Qwen2ForCausalLM"
@@ -34,13 +34,13 @@
34
  "forced_bos_token_id": null,
35
  "forced_eos_token_id": null,
36
  "hidden_act": "silu",
37
- "hidden_size": 1536,
38
  "id2label": {
39
  "0": "LABEL_0",
40
  "1": "LABEL_1"
41
  },
42
  "initializer_range": 0.02,
43
- "intermediate_size": 8960,
44
  "is_decoder": false,
45
  "is_encoder_decoder": false,
46
  "label2id": {
@@ -50,15 +50,15 @@
50
  "length_penalty": 1.0,
51
  "max_length": 20,
52
  "max_position_embeddings": 32768,
53
- "max_window_layers": 21,
54
  "min_length": 0,
55
  "model_type": "qwen2",
56
  "no_repeat_ngram_size": 0,
57
- "num_attention_heads": 12,
58
  "num_beam_groups": 1,
59
  "num_beams": 1,
60
- "num_hidden_layers": 28,
61
- "num_key_value_heads": 2,
62
  "num_return_sequences": 1,
63
  "output_attentions": false,
64
  "output_hidden_states": false,
@@ -81,7 +81,7 @@
81
  "temperature": 1.0,
82
  "tf_legacy_loss": false,
83
  "tie_encoder_decoder": false,
84
- "tie_word_embeddings": true,
85
  "tokenizer_class": null,
86
  "top_k": 50,
87
  "top_p": 1.0,
@@ -91,13 +91,13 @@
91
  "use_bfloat16": false,
92
  "use_cache": true,
93
  "use_sliding_window": false,
94
- "vocab_size": 151936
95
  },
96
  "model_type": "ovis",
97
  "multimodal_max_length": 2048,
98
- "torch_dtype": "float32",
99
- "transformers_version": "4.51.3",
100
- "use_cache": true,
101
  "visual_tokenizer_config": {
102
  "_attn_implementation_autoset": true,
103
  "_name_or_path": "",
@@ -105,7 +105,7 @@
105
  "architectures": null,
106
  "backbone_config": {
107
  "_attn_implementation_autoset": true,
108
- "_name_or_path": "apple/aimv2-large-patch14-448",
109
  "add_cross_attention": false,
110
  "architectures": [
111
  "AIMv2Model"
@@ -131,13 +131,13 @@
131
  "finetuning_task": null,
132
  "forced_bos_token_id": null,
133
  "forced_eos_token_id": null,
134
- "hidden_size": 1024,
135
  "id2label": {
136
  "0": "LABEL_0",
137
  "1": "LABEL_1"
138
  },
139
  "image_size": 448,
140
- "intermediate_size": 2816,
141
  "is_decoder": false,
142
  "is_encoder_decoder": false,
143
  "label2id": {
@@ -149,7 +149,7 @@
149
  "min_length": 0,
150
  "model_type": "aimv2",
151
  "no_repeat_ngram_size": 0,
152
- "num_attention_heads": 8,
153
  "num_beam_groups": 1,
154
  "num_beams": 1,
155
  "num_channels": 3,
@@ -180,7 +180,7 @@
180
  "tokenizer_class": null,
181
  "top_k": 50,
182
  "top_p": 1.0,
183
- "torch_dtype": "bfloat16",
184
  "torchscript": false,
185
  "typical_p": 1.0,
186
  "use_bfloat16": false,
 
4
  ],
5
  "auto_map": {
6
  "AutoConfig": "configuration_ovis.OvisConfig",
7
+ "AutoModelForCausalLM": "modeling_ovis.Ovis"
8
  },
9
  "conversation_formatter_class": "QwenConversationFormatter",
10
  "disable_tie_weight": false,
11
+ "hidden_size": 5120,
12
  "llm_attn_implementation": "flash_attention_2",
13
  "llm_config": {
14
  "_attn_implementation_autoset": true,
15
+ "_name_or_path": "Qwen/Qwen2.5-32B-Instruct",
16
  "add_cross_attention": false,
17
  "architectures": [
18
  "Qwen2ForCausalLM"
 
34
  "forced_bos_token_id": null,
35
  "forced_eos_token_id": null,
36
  "hidden_act": "silu",
37
+ "hidden_size": 5120,
38
  "id2label": {
39
  "0": "LABEL_0",
40
  "1": "LABEL_1"
41
  },
42
  "initializer_range": 0.02,
43
+ "intermediate_size": 27648,
44
  "is_decoder": false,
45
  "is_encoder_decoder": false,
46
  "label2id": {
 
50
  "length_penalty": 1.0,
51
  "max_length": 20,
52
  "max_position_embeddings": 32768,
53
+ "max_window_layers": 70,
54
  "min_length": 0,
55
  "model_type": "qwen2",
56
  "no_repeat_ngram_size": 0,
57
+ "num_attention_heads": 40,
58
  "num_beam_groups": 1,
59
  "num_beams": 1,
60
+ "num_hidden_layers": 64,
61
+ "num_key_value_heads": 8,
62
  "num_return_sequences": 1,
63
  "output_attentions": false,
64
  "output_hidden_states": false,
 
81
  "temperature": 1.0,
82
  "tf_legacy_loss": false,
83
  "tie_encoder_decoder": false,
84
+ "tie_word_embeddings": false,
85
  "tokenizer_class": null,
86
  "top_k": 50,
87
  "top_p": 1.0,
 
91
  "use_bfloat16": false,
92
  "use_cache": true,
93
  "use_sliding_window": false,
94
+ "vocab_size": 152064
95
  },
96
  "model_type": "ovis",
97
  "multimodal_max_length": 2048,
98
+ "torch_dtype": "float16",
99
+ "transformers_version": "4.46.2",
100
+ "use_cache": false,
101
  "visual_tokenizer_config": {
102
  "_attn_implementation_autoset": true,
103
  "_name_or_path": "",
 
105
  "architectures": null,
106
  "backbone_config": {
107
  "_attn_implementation_autoset": true,
108
+ "_name_or_path": "apple/aimv2-1B-patch14-448",
109
  "add_cross_attention": false,
110
  "architectures": [
111
  "AIMv2Model"
 
131
  "finetuning_task": null,
132
  "forced_bos_token_id": null,
133
  "forced_eos_token_id": null,
134
+ "hidden_size": 2048,
135
  "id2label": {
136
  "0": "LABEL_0",
137
  "1": "LABEL_1"
138
  },
139
  "image_size": 448,
140
+ "intermediate_size": 5632,
141
  "is_decoder": false,
142
  "is_encoder_decoder": false,
143
  "label2id": {
 
149
  "min_length": 0,
150
  "model_type": "aimv2",
151
  "no_repeat_ngram_size": 0,
152
+ "num_attention_heads": 16,
153
  "num_beam_groups": 1,
154
  "num_beams": 1,
155
  "num_channels": 3,
 
180
  "tokenizer_class": null,
181
  "top_k": 50,
182
  "top_p": 1.0,
183
+ "torch_dtype": "float32",
184
  "torchscript": false,
185
  "typical_p": 1.0,
186
  "use_bfloat16": false,