JadeRay-42 commited on
Commit
237e0b6
·
verified ·
1 Parent(s): b295f38

Upload 3 files

Browse files
Files changed (3) hide show
  1. config.json +85 -0
  2. preprocessor_config.json +22 -0
  3. pytorch_model.bin +3 -0
config.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "activation_dropout": 0.0,
3
+ "activation_function": "relu",
4
+ "angle_loss_coefficient": 1,
5
+ "architectures": [
6
+ "Mono3DVGv2ForSingleObjectDetection"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "auxiliary_loss": false,
10
+ "backbone": "swin_large_patch4_window7_224",
11
+ "backbone_config": null,
12
+ "bbox_cost": 5,
13
+ "bbox_loss_coefficient": 5,
14
+ "center3d_cost": 10,
15
+ "center3d_loss_coefficient": 10,
16
+ "class_cost": 2,
17
+ "cls_loss_coefficient": 2,
18
+ "d_model": 256,
19
+ "decoder_attention_heads": 8,
20
+ "decoder_depth_residual": false,
21
+ "decoder_ffn_dim": 256,
22
+ "decoder_layers": 1,
23
+ "decoder_n_points": 4,
24
+ "decoder_self_attn": false,
25
+ "decoder_text_residual": false,
26
+ "depth_loss_coefficient": 1,
27
+ "depth_map_loss_coefficient": 1,
28
+ "depth_max": 60.0,
29
+ "depth_min": 0.001,
30
+ "dim_loss_coefficient": 1,
31
+ "disable_custom_kernels": false,
32
+ "dropout": 0.1,
33
+ "encoder_attention_heads": 8,
34
+ "encoder_ffn_dim": 256,
35
+ "encoder_layers": 3,
36
+ "encoder_n_points": 4,
37
+ "focal_alpha": 0.25,
38
+ "freeze_backbone": true,
39
+ "giou_cost": 2,
40
+ "giou_loss_coefficient": 2,
41
+ "id2label": {
42
+ "0": "pedestrian",
43
+ "1": "car",
44
+ "2": "cyclist",
45
+ "3": "van",
46
+ "4": "truck",
47
+ "5": "tram",
48
+ "6": "bus",
49
+ "7": "person_sitting",
50
+ "8": "motorcyclist"
51
+ },
52
+ "init_box": false,
53
+ "init_std": 0.02,
54
+ "init_xavier_std": 1.0,
55
+ "is_encoder_decoder": true,
56
+ "label2id": {
57
+ "bus": 6,
58
+ "car": 1,
59
+ "cyclist": 2,
60
+ "motorcyclist": 8,
61
+ "pedestrian": 0,
62
+ "person_sitting": 7,
63
+ "tram": 5,
64
+ "truck": 4,
65
+ "van": 3
66
+ },
67
+ "model_type": "mono3dvgv2",
68
+ "num_channels": 3,
69
+ "num_depth_bins": 80,
70
+ "num_feature_levels": 4,
71
+ "num_queries": 1,
72
+ "num_text_output_layers": 1,
73
+ "position_embedding_type": "sine",
74
+ "pretrained_backbone_path": "pretrained-models/swin_large_patch4_window7_224/model.safetensors",
75
+ "text_encoder_type": "pretrained-models/roberta-base",
76
+ "torch_dtype": "float32",
77
+ "transformers_version": "4.44.0",
78
+ "two_stage": false,
79
+ "use_dab": true,
80
+ "use_pretrained_backbone": true,
81
+ "use_text_guided_adapter": true,
82
+ "use_timm_backbone": true,
83
+ "vl_encoder_type": "fusion",
84
+ "with_box_refine": true
85
+ }
preprocessor_config.json ADDED
@@ -0,0 +1,22 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "do_normalize": true,
3
+ "do_rescale": true,
4
+ "do_resize": true,
5
+ "image_mean": [
6
+ 0.485,
7
+ 0.456,
8
+ 0.406
9
+ ],
10
+ "image_processor_type": "Mono3DVGImageProcessor",
11
+ "image_std": [
12
+ 0.229,
13
+ 0.224,
14
+ 0.225
15
+ ],
16
+ "resample": 2,
17
+ "rescale_factor": 0.00392156862745098,
18
+ "size": {
19
+ "height": 384,
20
+ "width": 1280
21
+ }
22
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:eeae527ec76a06e746b913e8e4b4a984107567a03d67494cacdae822835aba3b
3
+ size 1324513302