ZrH42 commited on
Commit
165730b
·
1 Parent(s): da3cfa5
config.json ADDED
@@ -0,0 +1,66 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "aligner_config": {
3
+ "cls": "MlpProjector",
4
+ "model_type": "aligner",
5
+ "params": {
6
+ "depth": 2,
7
+ "input_dim": 1024,
8
+ "n_embed": 2048,
9
+ "projector_type": "mlp_gelu"
10
+ }
11
+ },
12
+ "architectures": [
13
+ "MultiModalityCausalLM"
14
+ ],
15
+ "gen_aligner_config": {
16
+ "cls": "MlpProjector",
17
+ "model_type": "gen_aligner",
18
+ "params": {
19
+ "depth": 2,
20
+ "input_dim": 8,
21
+ "n_embed": 2048,
22
+ "projector_type": "mlp_gelu"
23
+ }
24
+ },
25
+ "gen_head_config": {
26
+ "cls": "vision_head",
27
+ "model_type": "gen_head",
28
+ "params": {
29
+ "image_token_embed": 2048,
30
+ "image_token_size": 16384,
31
+ "n_embed": 2048
32
+ }
33
+ },
34
+ "gen_vision_config": {
35
+ "cls": "VQ-16",
36
+ "model_type": "gen_vision",
37
+ "params": {
38
+ "image_token_size": 16384,
39
+ "n_embed": 8
40
+ }
41
+ },
42
+ "language_config": {
43
+ "hidden_size": 2048,
44
+ "intermediate_size": 5632,
45
+ "max_position_embeddings": 16384,
46
+ "model_type": "llama",
47
+ "num_attention_heads": 16,
48
+ "num_hidden_layers": 24,
49
+ "num_key_value_heads": 16,
50
+ "torch_dtype": "bfloat16",
51
+ "vocab_size": 102400
52
+ },
53
+ "model_type": "multi_modality",
54
+ "torch_dtype": "bfloat16",
55
+ "transformers_version": "4.33.1",
56
+ "vision_config": {
57
+ "cls": "CLIPVisionTower",
58
+ "model_type": "vision",
59
+ "params": {
60
+ "image_size": 384,
61
+ "model_name": "siglip_large_patch16_384",
62
+ "select_feature": "same",
63
+ "select_layer": -1
64
+ }
65
+ }
66
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "background_color": [
3
+ 127,
4
+ 127,
5
+ 127
6
+ ],
7
+ "do_normalize": true,
8
+ "image_mean": [
9
+ 0.5,
10
+ 0.5,
11
+ 0.5
12
+ ],
13
+ "image_processor_type": "VLMImageProcessor",
14
+ "image_size": 384,
15
+ "image_std": [
16
+ 0.5,
17
+ 0.5,
18
+ 0.5
19
+ ],
20
+ "min_size": 14,
21
+ "processor_class": "VLChatProcessor",
22
+ "rescale_factor": 0.00392156862745098
23
+ }
processor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_special_token": false,
3
+ "ignore_id": -100,
4
+ "image_tag": "<image_placeholder>",
5
+ "mask_prompt": true,
6
+ "num_image_tokens": 576,
7
+ "processor_class": "VLChatProcessor",
8
+ "sft_format": "deepseek"
9
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<image_placeholder>",
4
+ "<patch_placeholder>",
5
+ "<|ref|>",
6
+ "<|/ref|>",
7
+ "<|det|>",
8
+ "<|/det|>",
9
+ "<|grounding|>",
10
+ "<|User|>",
11
+ "<|Assistant|>"
12
+ ],
13
+ "bos_token": "<|begin▁of▁sentence|>",
14
+ "eos_token": "<|end▁of▁sentence|>",
15
+ "pad_token": "<|▁pad▁|>"
16
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|begin▁of▁sentence|>",
3
+ "clean_up_tokenization_spaces": false,
4
+ "eos_token": "<|end▁of▁sentence|>",
5
+ "model_max_length": 16384,
6
+ "pad_token": null,
7
+ "tokenizer_class": "LlamaTokenizer",
8
+ "unk_token": null,
9
+ "use_default_system_prompt": true
10
+ }
vae/config.yaml ADDED
@@ -0,0 +1,44 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ model:
2
+ base_learning_rate: 4.5e-06
3
+ target: ldm.models.autoencoder.AutoencoderKL
4
+ params:
5
+ monitor: val/rec_loss
6
+ embed_dim: 16
7
+ lossconfig:
8
+ target: ldm.modules.losses.LPIPSWithDiscriminator
9
+ params:
10
+ disc_start: 50001
11
+ kl_weight: 1.0e-06
12
+ disc_weight: 0.5
13
+ ddconfig:
14
+ double_z: true
15
+ z_channels: 16
16
+ resolution: 256
17
+ in_channels: 3
18
+ out_ch: 3
19
+ ch: 128
20
+ ch_mult:
21
+ - 1
22
+ - 1
23
+ - 2
24
+ - 2
25
+ - 4
26
+ num_res_blocks: 2
27
+ attn_resolutions:
28
+ - 16
29
+ dropout: 0.0
30
+ data:
31
+ target: main.DataModuleFromConfig
32
+ params:
33
+ batch_size: 6
34
+ wrap: true
35
+ train:
36
+ target: ldm.data.openimages.FullOpenImagesTrain
37
+ params:
38
+ size: 384
39
+ crop_size: 256
40
+ validation:
41
+ target: ldm.data.openimages.FullOpenImagesValidation
42
+ params:
43
+ size: 384
44
+ crop_size: 256
vae/kl-f16d16.ckpt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:800019757fb0942448c7bcc955cb60744a349cdc986e0879564b0a1fbe65f835
3
+ size 927775203
vocab.json ADDED
The diff for this file is too large to render. See raw diff