Transformers
PyTorch
English
bridgetower
anahita-b commited on
Commit
dec62c1
·
1 Parent(s): 020392c

Separate config into text and vision configs

Browse files
Files changed (1) hide show
  1. config.json +37 -21
config.json CHANGED
@@ -1,9 +1,5 @@
1
  {
2
  "drop_rate":0.1,
3
- "freeze_RoBERTa":false,
4
- "freeze_ViT":false,
5
- "freeze_layer_count_roberta":false,
6
- "freeze_layer_count_vit":false,
7
  "head_hidden_scale":2,
8
  "hidden_act":"gelu",
9
  "hidden_size":768,
@@ -18,23 +14,43 @@
18
  "mlp_ratio":4,
19
  "num_attention_heads":12,
20
  "num_hidden_layers":6,
21
- "resolution_before":224,
22
  "stop_gradient":false,
23
  "tie_word_embeddings":false,
24
- "tokenizer":"roberta-base",
25
- "unfreeze_RoBERTa_attention":false,
26
- "unfreeze_RoBERTa_embeddings":false,
27
- "unfreeze_RoBERTa_encoder":false,
28
- "unfreeze_RoBERTa_layernorm":false,
29
- "unfreeze_ViT_attention":false,
30
- "unfreeze_ViT_layernorm":false,
31
- "vit_embed_dim":512,
32
- "vit_layers":12,
33
- "vit_layernorm_init_from_vit":false,
34
- "vit_layernorm_shared":true,
35
- "vit_patch_size":16,
36
- "vit_remove_last":false,
37
- "vit_transformer_width":512,
38
- "vit_width":768,
39
- "vocab_size":50265
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
40
  }
 
1
  {
2
  "drop_rate":0.1,
 
 
 
 
3
  "head_hidden_scale":2,
4
  "hidden_act":"gelu",
5
  "hidden_size":768,
 
14
  "mlp_ratio":4,
15
  "num_attention_heads":12,
16
  "num_hidden_layers":6,
 
17
  "stop_gradient":false,
18
  "tie_word_embeddings":false,
19
+ "vocab_size":50265,
20
+ "text_config_dict": null,
21
+ "text_config":{
22
+ "architectures": ["BridgeTowerTextModel"],
23
+ "classifier_dropout": null,
24
+ "vocab_size": 50265,
25
+ "hidden_size": 768,
26
+ "num_hidden_layers": 12,
27
+ "num_attention_heads": 12,
28
+ "intermediate_size": 3072,
29
+ "hidden_act": "gelu",
30
+ "hidden_dropout_prob": 0.1,
31
+ "attention_probs_dropout_prob": 0.1,
32
+ "max_position_embeddings": 514,
33
+ "type_vocab_size": 1,
34
+ "initializer_range": 0.02,
35
+ "layer_norm_eps": 1e-05,
36
+ "pad_token_id": 1,
37
+ "bos_token_id": 0,
38
+ "eos_token_id": 2,
39
+ "position_embedding_type": "absolute",
40
+ "use_cache": true
41
+ },
42
+ "vision_config_dict": null,
43
+ "vision_config":{
44
+ "architectures": ["BridgeTowerVisionModel"],
45
+ "embed_dim": 512,
46
+ "input_resolution": 224,
47
+ "width": 768,
48
+ "layers": 12,
49
+ "patch_size": 16,
50
+ "transformer_width": 512,
51
+ "output_resolution": 288,
52
+ "stop_gradient": false,
53
+ "vit_layernorm_shared": true,
54
+ "vit_remove_last": false
55
+ }
56
  }