cuijh26 commited on
Commit
ae49e8a
·
verified ·
1 Parent(s): 31281a4

Delete WAM-Flow-NAVSIM

Browse files
WAM-Flow-NAVSIM/config.json DELETED
@@ -1,67 +0,0 @@
1
- {
2
- "_name_or_path": "/cache/models/LucasJinWang-FUDOKI",
3
- "aligner_config": {
4
- "cls": "MlpProjector",
5
- "model_type": "aligner",
6
- "params": {
7
- "depth": 2,
8
- "input_dim": 1024,
9
- "n_embed": 2048,
10
- "projector_type": "mlp_gelu"
11
- }
12
- },
13
- "architectures": [
14
- "MultiModalityCausalLM"
15
- ],
16
- "gen_aligner_config": {
17
- "cls": "MlpProjector",
18
- "model_type": "gen_aligner",
19
- "params": {
20
- "depth": 2,
21
- "input_dim": 8,
22
- "n_embed": 2048,
23
- "projector_type": "mlp_gelu"
24
- }
25
- },
26
- "gen_head_config": {
27
- "cls": "vision_head",
28
- "model_type": "gen_head",
29
- "params": {
30
- "image_token_embed": 2048,
31
- "image_token_size": 16384,
32
- "n_embed": 2048
33
- }
34
- },
35
- "gen_vision_config": {
36
- "cls": "VQ-16",
37
- "model_type": "gen_vision",
38
- "params": {
39
- "image_token_size": 16384,
40
- "n_embed": 8
41
- }
42
- },
43
- "language_config": {
44
- "hidden_size": 2048,
45
- "intermediate_size": 5632,
46
- "max_position_embeddings": 16384,
47
- "model_type": "llama",
48
- "num_attention_heads": 16,
49
- "num_hidden_layers": 24,
50
- "num_key_value_heads": 16,
51
- "torch_dtype": "bfloat16",
52
- "vocab_size": 122401
53
- },
54
- "model_type": "multi_modality",
55
- "torch_dtype": "float32",
56
- "transformers_version": "4.42.4",
57
- "vision_config": {
58
- "cls": "CLIPVisionTower",
59
- "model_type": "vision",
60
- "params": {
61
- "image_size": 384,
62
- "model_name": "siglip_large_patch16_384",
63
- "select_feature": "same",
64
- "select_layer": -1
65
- }
66
- }
67
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
WAM-Flow-NAVSIM/model.safetensors DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b06580ad28d20eeade8c103c70cb60ac8d538bc20d995684aedfa2674babf02
3
- size 8684997012