aagarwalRT commited on
Commit
6b869b3
·
1 Parent(s): a252638

Llava Weights

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
added_tokens.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "<image>": 32000,
3
+ "<pad>": 32001
4
+ }
mlc-chat-config.json ADDED
@@ -0,0 +1,175 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "0.1.0",
3
+ "model_type": "llava",
4
+ "quantization": "q4f16_1",
5
+ "model_config": {
6
+ "image_token_index": 32000,
7
+ "text_config": {
8
+ "hidden_size": 4096,
9
+ "intermediate_size": 11008,
10
+ "num_attention_heads": 32,
11
+ "num_hidden_layers": 32,
12
+ "rms_norm_eps": 1e-05,
13
+ "vocab_size": 32000,
14
+ "position_embedding_base": 10000.0,
15
+ "context_window_size": 4096,
16
+ "prefill_chunk_size": 2048,
17
+ "num_key_value_heads": 32,
18
+ "head_dim": 128,
19
+ "tensor_parallel_shards": 1,
20
+ "max_batch_size": 1,
21
+ "kwargs": {
22
+ "_name_or_path": "lmsys/vicuna-7b-v1.5",
23
+ "architectures": [
24
+ "LlamaForCausalLM"
25
+ ],
26
+ "model_type": "llama",
27
+ "torch_dtype": "float16",
28
+ "hidden_act": "silu",
29
+ "initializer_range": 0.02,
30
+ "pretraining_tp": 1,
31
+ "use_cache": true,
32
+ "rope_scaling": null,
33
+ "attention_bias": false,
34
+ "attention_dropout": 0.0,
35
+ "return_dict": true,
36
+ "output_hidden_states": false,
37
+ "output_attentions": false,
38
+ "torchscript": false,
39
+ "use_bfloat16": false,
40
+ "tf_legacy_loss": false,
41
+ "pruned_heads": {},
42
+ "tie_word_embeddings": false,
43
+ "chunk_size_feed_forward": 0,
44
+ "is_encoder_decoder": false,
45
+ "is_decoder": false,
46
+ "cross_attention_hidden_size": null,
47
+ "add_cross_attention": false,
48
+ "tie_encoder_decoder": false,
49
+ "max_length": 20,
50
+ "min_length": 0,
51
+ "do_sample": false,
52
+ "early_stopping": false,
53
+ "num_beams": 1,
54
+ "num_beam_groups": 1,
55
+ "diversity_penalty": 0.0,
56
+ "temperature": 1.0,
57
+ "top_k": 50,
58
+ "top_p": 1.0,
59
+ "typical_p": 1.0,
60
+ "repetition_penalty": 1.0,
61
+ "length_penalty": 1.0,
62
+ "no_repeat_ngram_size": 0,
63
+ "encoder_no_repeat_ngram_size": 0,
64
+ "bad_words_ids": null,
65
+ "num_return_sequences": 1,
66
+ "output_scores": false,
67
+ "return_dict_in_generate": false,
68
+ "forced_bos_token_id": null,
69
+ "forced_eos_token_id": null,
70
+ "remove_invalid_values": false,
71
+ "exponential_decay_length_penalty": null,
72
+ "suppress_tokens": null,
73
+ "begin_suppress_tokens": null,
74
+ "finetuning_task": null,
75
+ "id2label": {
76
+ "0": "LABEL_0",
77
+ "1": "LABEL_1"
78
+ },
79
+ "label2id": {
80
+ "LABEL_0": 0,
81
+ "LABEL_1": 1
82
+ },
83
+ "tokenizer_class": null,
84
+ "prefix": null,
85
+ "bos_token_id": 1,
86
+ "pad_token_id": 0,
87
+ "eos_token_id": 2,
88
+ "sep_token_id": null,
89
+ "decoder_start_token_id": null,
90
+ "task_specific_params": null,
91
+ "problem_type": null,
92
+ "transformers_version": "4.37.2"
93
+ }
94
+ },
95
+ "vision_config": {
96
+ "hidden_size": 1024,
97
+ "image_size": 336,
98
+ "intermediate_size": 4096,
99
+ "num_attention_heads": 16,
100
+ "num_hidden_layers": 24,
101
+ "patch_size": 14,
102
+ "projection_dim": 768,
103
+ "vocab_size": 32000,
104
+ "num_channels": 3,
105
+ "layer_norm_eps": 1e-06,
106
+ "kwargs": {
107
+ "model_type": "clip_vision_model"
108
+ }
109
+ },
110
+ "vocab_size": 32064,
111
+ "context_window_size": 4096,
112
+ "sliding_window_size": -1,
113
+ "prefill_chunk_size": 2048,
114
+ "tensor_parallel_shards": 1,
115
+ "max_batch_size": 80,
116
+ "text_architecture": "LlamaForCausalLM"
117
+ },
118
+ "vocab_size": 32064,
119
+ "context_window_size": 4096,
120
+ "sliding_window_size": -1,
121
+ "prefill_chunk_size": 2048,
122
+ "attention_sink_size": -1,
123
+ "tensor_parallel_shards": 1,
124
+ "temperature": 1.0,
125
+ "presence_penalty": 0.0,
126
+ "frequency_penalty": 0.0,
127
+ "repetition_penalty": 1.0,
128
+ "top_p": 1.0,
129
+ "tokenizer_files": [
130
+ "tokenizer.model",
131
+ "tokenizer.json",
132
+ "added_tokens.json",
133
+ "tokenizer_config.json"
134
+ ],
135
+ "tokenizer_info": {
136
+ "token_postproc_method": "byte_fallback",
137
+ "prepend_space_in_encode": true,
138
+ "strip_space_in_decode": true
139
+ },
140
+ "conv_template": {
141
+ "name": "llava",
142
+ "system_template": "{system_message}",
143
+ "system_message": "\n",
144
+ "system_prefix_token_ids": [
145
+ 1
146
+ ],
147
+ "add_role_after_system_message": false,
148
+ "roles": {
149
+ "user": "USER",
150
+ "assistant": "ASSISTANT"
151
+ },
152
+ "role_templates": {
153
+ "user": "{user_message}",
154
+ "assistant": "{assistant_message}",
155
+ "tool": "{tool_message}"
156
+ },
157
+ "messages": [],
158
+ "seps": [
159
+ " "
160
+ ],
161
+ "role_content_sep": ": ",
162
+ "role_empty_sep": ":",
163
+ "stop_str": [
164
+ "</s>"
165
+ ],
166
+ "stop_token_ids": [
167
+ 2
168
+ ],
169
+ "function_string": "",
170
+ "use_function_calling": false
171
+ },
172
+ "pad_token_id": 32001,
173
+ "bos_token_id": 1,
174
+ "eos_token_id": 2
175
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d82c1921dfd29eefc53810f9e85c60e304450c768b0d88953479a1cfd453c9bb
3
+ size 65667072
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:537c1db2d9970d701ad6c9337356b1853cac33f8c98b02bf4862ddd4be9a124b
3
+ size 30760960
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f70d1bf5625a49a992d09f326ff5acfc18a4f8d4802328b96b8768b319cc43f
3
+ size 31989760
params_shard_100.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6be7068f0849b953048d5a6322839f639fdc7e25d2f944b4f772598ebb79163b
3
+ size 32661504
params_shard_101.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eff7d81dca799ffb14e384b3abf672bd761fa6f78e5abbe423b131540e741fbd
3
+ size 31989760
params_shard_102.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:043a577b34858f48d33d2c53fa6476eabd3f8429066ef7a85e90d9db6d92aa00
3
+ size 45088768
params_shard_103.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f794133a9c06e754519b2d86202cededdd2a4e7764895ffa98754ef9f64c05aa
3
+ size 25165824
params_shard_104.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b335ddb56f6d17ca0612df97559a65d7942491f4a98f298a9d6ab88b37d66155
3
+ size 22544384
params_shard_105.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc15ca71d32e81271e9084e6b1041d20a7a30173838fbbc0f035b4a6347c4aad
3
+ size 45088768
params_shard_106.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4f890f2dd843534e16a08ebed8cf06699a6f3ffcc0229964749a15ed190aa65f
3
+ size 25165824
params_shard_107.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f04c111ff47ec27aad04690afce9d6d94a311e1c669dafd3bf4e1eabc21c1c7f
3
+ size 32661504
params_shard_108.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff448b44347b2ac8144fb18a3693af6674eab72b94e1342eac1faa5f7ea7c0f8
3
+ size 31989760
params_shard_109.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5a87ab2ae2c346b29f2a2c8e8ec3a596848f0508cc628710a35057fafbdae96a
3
+ size 45088768
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6008202d5c5568b7425bdacc565928a5c587c893f98588a5f536ce1202650899
3
+ size 45088768
params_shard_110.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26d446d8c2a4d414b4c6705d39512870ae0b37ab9399194f0aca10adf9d45323
3
+ size 25165824
params_shard_111.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a0c6b6b6494b6e67ab7ef05fdf33db5e4313f7b498f4343134d0d076b7a6c2fc
3
+ size 22544384
params_shard_112.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c3feadd8571baeabe89e596df8387f65087d3e75a635c68a4f8b396888532c9
3
+ size 45088768
params_shard_113.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:adab41b68d378e3ce444183d350f65b2613cf642efaf5d36206853d38d1a07d4
3
+ size 25165824
params_shard_114.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1bc9f6f5e9424ddde172f8e00e2c040f7ca5dddf2a715f9bdd8ffb9ddab37fa3
3
+ size 32661504
params_shard_115.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:554228f1e29cc794fd82252daac551dda247e1f539938639f2721c9204e44aeb
3
+ size 31989760
params_shard_116.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bd6edbc5663f8c6eb103b36184ae7353275d5707ba4a9984efc444dc72753818
3
+ size 45088768
params_shard_117.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d80cb8e5eddee1c364bd19f0ee8250ffbfcdec9f21e32820992db6a97331d35
3
+ size 25165824
params_shard_118.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:940e4896458f8f235cc7f3903aec0e9a1a23bb2e1b0101e969d707704c119338
3
+ size 25165824
params_shard_119.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96e0946ba5854b9579c1da968cd053f3d3aefddae5a8b3498bb26a0d0601fa12
3
+ size 32579584
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65d3aecdccc4279de08b7c77ce5208adfac5ac0d13c7a677ac91bee3542d0c20
3
+ size 25165824
params_shard_120.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dfe3a6750c631c4b4c8d1c11a68ed60f1374452844228535d842c45f4d8aec4d
3
+ size 1048576
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d8e9eda1c13d17e21d74a03c2ba2ec2c910cc2275ba59ac4c599726873b65213
3
+ size 22544384
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:db8c15a6d20aa61603694eabd6026c9ff3876d37030f676762cfc7360fd7f45d
3
+ size 45088768
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a992ebaa99085a6d04760eeca6d16e4539978e594841286343284fe3a6065f3b
3
+ size 25165824
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc55cb5ef19652da9d6fca0c0d296eaa738b94a12f67bf9c324fd30ac7da074c
3
+ size 32661504
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a143d2d001a9d210578f37f1bb94327cdfc249093adb4fb1c8a735a3905c4073
3
+ size 31989760
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be7fa2b2186dc5da7f604c04d0245a2757c1a1afe75a73648a1c7edb69223ff6
3
+ size 45088768
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f962a6c30e2b0c6e026aa3692686588e5a8ec95f8f6e0ac22588c68f0ce444cd
3
+ size 25165824
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:718aebbdd7c94e988718ba710c56ecddcb4b33ca1234c6ad193b425045819d55
3
+ size 45088768
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84658ece31c303a290d3f38dc7093e6f07feb8878f434019bd80bf5093a5bdf2
3
+ size 22544384
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be80e29a5f6191ba156ef516cdb5719c80bb30e4ebb0a7a638e6528d4ae099e3
3
+ size 45088768
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a880dc8991adc85c3dade6cf58a844e8d5e80267e28259595f18d5316eb0f8ea
3
+ size 25165824
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:655abf2d898f482277a9e9f98ba02958d30b7f5a1627442530356123dd37f7f5
3
+ size 32661504
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d322c21acec65d76adbc1774aa77d730342fe5d7f99e08050d015d67f513f9b6
3
+ size 31989760
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9827236f11cfe0104f90dcbe2d3fa60e95d773ab16916bc294a78c37cf6dc936
3
+ size 45088768
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28886f0593a93862944dc65bbba129ace133ccca36e8a063e975e3b3a2f4b0bb
3
+ size 25165824
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9492468fea95db84faf525e0fee968973ff26091965c40a26734ba0022144cfc
3
+ size 22544384
params_shard_28.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ea4808eeed27fbf9617aabbfd41311dbc9092a750c3b3363091b225a528dd28
3
+ size 45088768
params_shard_29.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3097196f80522d78a3dd5977c5e3721d02ff357bbbc01d93bea3af1527fe71c0
3
+ size 25165824
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:614c610f9d3ffd92d6618b535add197b203101cf7f5be414e065bb4d8439bd1d
3
+ size 31014912
params_shard_30.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:970890afb8dcb6d1a2070cfbedd21cd365f380f04910df75bd97ef2eb2a4e956
3
+ size 32661504
params_shard_31.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:34430dbfaac542153821b1216f29fe4cd22f056b0e0ba0b54059f9145ba9b540
3
+ size 31989760