jirong commited on
Commit
2e72dd0
·
verified ·
1 Parent(s): c3d3db2

Upload folder using huggingface_hub

Browse files
added_tokens.json ADDED
The diff for this file is too large to render. See raw diff
 
config.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "/mnt/hdfs/public/jirong/codebase/RoboICL/models/pretrained_vl",
3
+ "architectures": [
4
+ "Qwen2ForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 1536,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 8960,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 28,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 12,
17
+ "num_hidden_layers": 28,
18
+ "num_key_value_heads": 2,
19
+ "rms_norm_eps": 1e-06,
20
+ "rope_scaling": null,
21
+ "rope_theta": 1000000.0,
22
+ "sliding_window": null,
23
+ "tie_word_embeddings": true,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.49.0.dev0",
26
+ "use_cache": true,
27
+ "use_mrope": false,
28
+ "use_sliding_window": false,
29
+ "vocab_size": 163419
30
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.49.0.dev0"
6
+ }
merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
preprocessor_config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_convert_rgb": true,
3
+ "do_normalize": false,
4
+ "do_rescale": true,
5
+ "do_resize": true,
6
+ "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
+ ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
12
+ "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
+ ],
17
+ "max_pixels": 1003520,
18
+ "merge_size": 1,
19
+ "min_pixels": 3136,
20
+ "patch_size": 16,
21
+ "resample": 3,
22
+ "rescale_factor": 0.00392156862745098,
23
+ "size": {
24
+ "max_pixels": 1003520,
25
+ "min_pixels": 3136
26
+ },
27
+ "temporal_patch_size": 1
28
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45e8a5d2f7909b688daca5b58cc67527884111998cce4ebbf7ae69139173fbe9
3
+ size 3122749946
special_tokens_map.json ADDED
@@ -0,0 +1,118 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|frame_0|>",
4
+ "<|frame_1|>",
5
+ "<|frame_2|>",
6
+ "<|frame_3|>",
7
+ "<|frame_4|>",
8
+ "<|frame_5|>",
9
+ "<|frame_6|>",
10
+ "<|frame_7|>",
11
+ "<|frame_8|>",
12
+ "<|frame_9|>",
13
+ "<|frame_10|>",
14
+ "<|frame_11|>",
15
+ "<|frame_12|>",
16
+ "<|frame_13|>",
17
+ "<|frame_14|>",
18
+ "<|frame_15|>",
19
+ "<|frame_16|>",
20
+ "<|frame_17|>",
21
+ "<|frame_18|>",
22
+ "<|frame_19|>",
23
+ "<|frame_20|>",
24
+ "<|frame_21|>",
25
+ "<|frame_22|>",
26
+ "<|frame_23|>",
27
+ "<|frame_24|>",
28
+ "<|frame_25|>",
29
+ "<|frame_26|>",
30
+ "<|frame_27|>",
31
+ "<|frame_28|>",
32
+ "<|frame_29|>",
33
+ "<|frame_30|>",
34
+ "<|frame_31|>",
35
+ "<|frame_32|>",
36
+ "<|frame_33|>",
37
+ "<|frame_34|>",
38
+ "<|frame_35|>",
39
+ "<|frame_36|>",
40
+ "<|frame_37|>",
41
+ "<|frame_38|>",
42
+ "<|frame_39|>",
43
+ "<|frame_40|>",
44
+ "<|frame_41|>",
45
+ "<|frame_42|>",
46
+ "<|frame_43|>",
47
+ "<|frame_44|>",
48
+ "<|frame_45|>",
49
+ "<|frame_46|>",
50
+ "<|frame_47|>",
51
+ "<|frame_48|>",
52
+ "<|frame_49|>",
53
+ "<|frame_50|>",
54
+ "<|frame_51|>",
55
+ "<|frame_52|>",
56
+ "<|frame_53|>",
57
+ "<|frame_54|>",
58
+ "<|frame_55|>",
59
+ "<|frame_56|>",
60
+ "<|frame_57|>",
61
+ "<|frame_58|>",
62
+ "<|frame_59|>",
63
+ "<|frame_60|>",
64
+ "<|frame_61|>",
65
+ "<|frame_62|>",
66
+ "<|frame_63|>",
67
+ "<|frame_64|>",
68
+ "<|frame_65|>",
69
+ "<|frame_66|>",
70
+ "<|frame_67|>",
71
+ "<|frame_68|>",
72
+ "<|frame_69|>",
73
+ "<|frame_70|>",
74
+ "<|frame_71|>",
75
+ "<|frame_72|>",
76
+ "<|frame_73|>",
77
+ "<|frame_74|>",
78
+ "<|frame_75|>",
79
+ "<|frame_76|>",
80
+ "<|frame_77|>",
81
+ "<|frame_78|>",
82
+ "<|frame_79|>",
83
+ "<|frame_80|>",
84
+ "<|frame_81|>",
85
+ "<|frame_82|>",
86
+ "<|frame_83|>",
87
+ "<|frame_84|>",
88
+ "<|frame_85|>",
89
+ "<|frame_86|>",
90
+ "<|frame_87|>",
91
+ "<|frame_88|>",
92
+ "<|frame_89|>",
93
+ "<|frame_90|>",
94
+ "<|frame_91|>",
95
+ "<|frame_92|>",
96
+ "<|frame_93|>",
97
+ "<|frame_94|>",
98
+ "<|frame_95|>",
99
+ "<|frame_96|>",
100
+ "<|frame_97|>",
101
+ "<|frame_98|>",
102
+ "<|frame_99|>"
103
+ ],
104
+ "eos_token": {
105
+ "content": "<|endoftext|>",
106
+ "lstrip": false,
107
+ "normalized": false,
108
+ "rstrip": false,
109
+ "single_word": false
110
+ },
111
+ "pad_token": {
112
+ "content": "<|endoftext|>",
113
+ "lstrip": false,
114
+ "normalized": false,
115
+ "rstrip": false,
116
+ "single_word": false
117
+ }
118
+ }
tokenizer_config.json ADDED
The diff for this file is too large to render. See raw diff
 
vocab.json ADDED
The diff for this file is too large to render. See raw diff