File size: 1,266 Bytes
d306695
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
{
  "additional_vocab_size": 40,
  "architectures": [
    "SparseModernVBertMLM"
  ],
  "auto_map": {
    "AutoConfig": "configuration_modernvbert.ModernVBertConfig",
    "AutoModel": "modeling_modernvbert.ModernVBertModel",
    "AutoModelForMaskedLM": "modeling_modernvbert.ModernVBertForMaskedLM"
  },
  "dtype": "float32",
  "freeze_config": null,
  "hidden_size": 768,
  "image_token_id": 50407,
  "initializer_range": 0.02,
  "max_position_embeddings": 8192,
  "model_type": "modernvbert",
  "neftune_noise_alpha": 0.0,
  "output_attentions": false,
  "pixel_shuffle_factor": 4,
  "qk_layer_norms": false,
  "scale_factor": 4,
  "text_config": {
    "hidden_size": 768,
    "intermediate_size": 1152,
    "mlp_bias": false,
    "model_type": "modernvbert_text",
    "num_hidden_layers": 22,
    "text_model_name": "jhu-clsp/ettin-encoder-150m",
    "vocab_size": 50368
  },
  "tie_word_embeddings": false,
  "transformers_version": null,
  "use_cache": true,
  "use_resampler": false,
  "vision_config": {
    "embed_dim": 768,
    "image_size": 512,
    "intermediate_size": 3072,
    "model_type": "modernvbert_vision",
    "num_hidden_layers": 12,
    "patch_size": 16,
    "vision_model_name": "google/siglip2-base-patch16-512"
  },
  "vocab_size": 50368
}