gogoduan commited on
Commit
ecddf57
·
verified ·
1 Parent(s): e5d2b62

Upload folder using huggingface_hub

Browse files
config.json ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "ckpts/Janus-Pro-1B",
3
+ "aligner_config": {
4
+ "cls": "MlpProjector",
5
+ "model_type": "aligner",
6
+ "params": {
7
+ "depth": 2,
8
+ "input_dim": 1024,
9
+ "n_embed": 2048,
10
+ "projector_type": "mlp_gelu"
11
+ }
12
+ },
13
+ "architectures": [
14
+ "MultiModalityCausalLM"
15
+ ],
16
+ "gen_aligner_config": {
17
+ "cls": "MlpProjector",
18
+ "model_type": "gen_aligner",
19
+ "params": {
20
+ "depth": 2,
21
+ "input_dim": 8,
22
+ "n_embed": 2048,
23
+ "projector_type": "mlp_gelu"
24
+ }
25
+ },
26
+ "gen_head_config": {
27
+ "cls": "vision_head",
28
+ "model_type": "gen_head",
29
+ "params": {
30
+ "image_token_embed": 2048,
31
+ "image_token_size": 16384,
32
+ "n_embed": 2048
33
+ }
34
+ },
35
+ "gen_vision_config": {
36
+ "cls": "VQ-16",
37
+ "model_type": "gen_vision",
38
+ "params": {
39
+ "image_token_size": 16384,
40
+ "n_embed": 8
41
+ }
42
+ },
43
+ "language_config": {
44
+ "_attn_implementation_autoset": true,
45
+ "hidden_size": 2048,
46
+ "intermediate_size": 5632,
47
+ "max_position_embeddings": 16384,
48
+ "model_type": "llama",
49
+ "num_attention_heads": 16,
50
+ "num_hidden_layers": 24,
51
+ "num_key_value_heads": 16,
52
+ "torch_dtype": "bfloat16",
53
+ "vocab_size": 102400
54
+ },
55
+ "model_type": "multi_modality",
56
+ "torch_dtype": "float16",
57
+ "transformers_version": "4.49.0",
58
+ "vision_config": {
59
+ "cls": "CLIPVisionTower",
60
+ "model_type": "vision",
61
+ "params": {
62
+ "image_size": 384,
63
+ "model_name": "siglip_large_patch16_384",
64
+ "select_feature": "same",
65
+ "select_layer": -1
66
+ }
67
+ }
68
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a3bf43dd1f41e2291452b3ca09f9571a3825225d779d8a0a5d3051f4e896823
3
+ size 4178705502
preprocessor_config.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "background_color": [
3
+ 127,
4
+ 127,
5
+ 127
6
+ ],
7
+ "do_normalize": true,
8
+ "image_mean": [
9
+ 0.5,
10
+ 0.5,
11
+ 0.5
12
+ ],
13
+ "image_processor_type": "VLMImageProcessor",
14
+ "image_size": 384,
15
+ "image_std": [
16
+ 0.5,
17
+ 0.5,
18
+ 0.5
19
+ ],
20
+ "min_size": 14,
21
+ "processor_class": "VLChatProcessor",
22
+ "rescale_factor": 0.00392156862745098
23
+ }
processor_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_special_token": false,
3
+ "ignore_id": -100,
4
+ "image_tag": "<image_placeholder>",
5
+ "mask_prompt": true,
6
+ "num_image_tokens": 576,
7
+ "processor_class": "VLChatProcessor",
8
+ "sft_format": "deepseek"
9
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<image_placeholder>",
4
+ "<patch_placeholder>",
5
+ "<|ref|>",
6
+ "<|/ref|>",
7
+ "<|det|>",
8
+ "<|/det|>",
9
+ "<|grounding|>",
10
+ "<|User|>",
11
+ "<|Assistant|>"
12
+ ],
13
+ "bos_token": "<|begin▁of▁sentence|>",
14
+ "eos_token": "<|end▁of▁sentence|>",
15
+ "pad_token": "<|▁pad▁|>"
16
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<|begin▁of▁sentence|>",
3
+ "clean_up_tokenization_spaces": false,
4
+ "eos_token": "<|end▁of▁sentence|>",
5
+ "model_max_length": 16384,
6
+ "pad_token": null,
7
+ "tokenizer_class": "LlamaTokenizer",
8
+ "unk_token": null,
9
+ "use_default_system_prompt": true
10
+ }