lewtun HF Staff commited on
Commit
a63dc40
·
1 Parent(s): 336b6ab

Upload model

Browse files
Files changed (2) hide show
  1. config.json +11 -11
  2. pytorch_model.bin +2 -2
config.json CHANGED
@@ -9,8 +9,8 @@
9
  "logit_scale_init_value": 2.6592,
10
  "model_type": "groupvit",
11
  "output_segmentation": false,
12
- "projection_dim": 16,
13
- "projection_intermediate_dim": 16,
14
  "text_config": {
15
  "_name_or_path": "",
16
  "add_cross_attention": false,
@@ -32,14 +32,14 @@
32
  "forced_bos_token_id": null,
33
  "forced_eos_token_id": null,
34
  "hidden_act": "quick_gelu",
35
- "hidden_size": 8,
36
  "id2label": {
37
  "0": "LABEL_0",
38
  "1": "LABEL_1"
39
  },
40
  "initializer_factor": 1.0,
41
  "initializer_range": 0.02,
42
- "intermediate_size": 16,
43
  "is_decoder": false,
44
  "is_encoder_decoder": false,
45
  "label2id": {
@@ -49,11 +49,11 @@
49
  "layer_norm_eps": 1e-05,
50
  "length_penalty": 1.0,
51
  "max_length": 20,
52
- "max_position_embeddings": 77,
53
  "min_length": 0,
54
  "model_type": "groupvit_text_model",
55
  "no_repeat_ngram_size": 0,
56
- "num_attention_heads": 2,
57
  "num_beam_groups": 1,
58
  "num_beams": 1,
59
  "num_hidden_layers": 5,
@@ -83,7 +83,7 @@
83
  "transformers_version": "4.22.0.dev0",
84
  "typical_p": 1.0,
85
  "use_bfloat16": false,
86
- "vocab_size": 49408
87
  },
88
  "text_config_dict": null,
89
  "torch_dtype": "float32",
@@ -119,7 +119,7 @@
119
  "forced_bos_token_id": null,
120
  "forced_eos_token_id": null,
121
  "hidden_act": "gelu",
122
- "hidden_size": 8,
123
  "id2label": {
124
  "0": "LABEL_0",
125
  "1": "LABEL_1"
@@ -127,7 +127,7 @@
127
  "image_size": 224,
128
  "initializer_factor": 1.0,
129
  "initializer_range": 0.02,
130
- "intermediate_size": 1536,
131
  "is_decoder": false,
132
  "is_encoder_decoder": false,
133
  "label2id": {
@@ -140,7 +140,7 @@
140
  "min_length": 0,
141
  "model_type": "groupvit_vision_model",
142
  "no_repeat_ngram_size": 0,
143
- "num_attention_heads": 2,
144
  "num_beam_groups": 1,
145
  "num_beams": 1,
146
  "num_channels": 3,
@@ -160,7 +160,7 @@
160
  "output_hidden_states": false,
161
  "output_scores": false,
162
  "pad_token_id": null,
163
- "patch_size": 16,
164
  "prefix": null,
165
  "problem_type": null,
166
  "pruned_heads": {},
 
9
  "logit_scale_init_value": 2.6592,
10
  "model_type": "groupvit",
11
  "output_segmentation": false,
12
+ "projection_dim": 64,
13
+ "projection_intermediate_dim": 4096,
14
  "text_config": {
15
  "_name_or_path": "",
16
  "add_cross_attention": false,
 
32
  "forced_bos_token_id": null,
33
  "forced_eos_token_id": null,
34
  "hidden_act": "quick_gelu",
35
+ "hidden_size": 32,
36
  "id2label": {
37
  "0": "LABEL_0",
38
  "1": "LABEL_1"
39
  },
40
  "initializer_factor": 1.0,
41
  "initializer_range": 0.02,
42
+ "intermediate_size": 37,
43
  "is_decoder": false,
44
  "is_encoder_decoder": false,
45
  "label2id": {
 
49
  "layer_norm_eps": 1e-05,
50
  "length_penalty": 1.0,
51
  "max_length": 20,
52
+ "max_position_embeddings": 512,
53
  "min_length": 0,
54
  "model_type": "groupvit_text_model",
55
  "no_repeat_ngram_size": 0,
56
+ "num_attention_heads": 4,
57
  "num_beam_groups": 1,
58
  "num_beams": 1,
59
  "num_hidden_layers": 5,
 
83
  "transformers_version": "4.22.0.dev0",
84
  "typical_p": 1.0,
85
  "use_bfloat16": false,
86
+ "vocab_size": 99
87
  },
88
  "text_config_dict": null,
89
  "torch_dtype": "float32",
 
119
  "forced_bos_token_id": null,
120
  "forced_eos_token_id": null,
121
  "hidden_act": "gelu",
122
+ "hidden_size": 32,
123
  "id2label": {
124
  "0": "LABEL_0",
125
  "1": "LABEL_1"
 
127
  "image_size": 224,
128
  "initializer_factor": 1.0,
129
  "initializer_range": 0.02,
130
+ "intermediate_size": 37,
131
  "is_decoder": false,
132
  "is_encoder_decoder": false,
133
  "label2id": {
 
140
  "min_length": 0,
141
  "model_type": "groupvit_vision_model",
142
  "no_repeat_ngram_size": 0,
143
+ "num_attention_heads": 4,
144
  "num_beam_groups": 1,
145
  "num_beams": 1,
146
  "num_channels": 3,
 
160
  "output_hidden_states": false,
161
  "output_scores": false,
162
  "pad_token_id": null,
163
+ "patch_size": 2,
164
  "prefix": null,
165
  "problem_type": null,
166
  "pruned_heads": {},
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c4dd422bd85ae1f96d5f8fad32ef13c453efc55b7c3a690c6ab14d1df82306c8
3
- size 3246131
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a25617c44bccd28450fc05e6cb26d64a22b0ca28fb71d454944be50916a87df5
3
+ size 5770035