stevenbucaille commited on
Commit
66a3535
·
verified ·
1 Parent(s): f9950b4

Add rf-detr-large model

Browse files
Files changed (2) hide show
  1. config.json +28 -29
  2. model.safetensors +2 -2
config.json CHANGED
@@ -13,30 +13,30 @@
13
  "drop_path_rate": 0.0,
14
  "hidden_act": "gelu",
15
  "hidden_dropout_prob": 0.0,
16
- "hidden_size": 768,
17
- "image_size": 518,
18
  "initializer_range": 0.02,
19
  "layer_norm_eps": 1e-06,
20
  "layerscale_value": 1.0,
21
  "mlp_ratio": 4,
22
  "model_type": "rf_detr_dinov2",
23
- "num_attention_heads": 12,
24
  "num_channels": 3,
25
  "num_hidden_layers": 12,
26
- "num_windows": 4,
27
  "out_features": [
28
- "stage2",
29
- "stage5",
30
- "stage8",
31
- "stage11"
32
  ],
33
  "out_indices": [
34
- 2,
35
- 5,
36
- 8,
37
- 11
38
  ],
39
- "patch_size": 14,
40
  "qkv_bias": true,
41
  "reshape_hidden_states": true,
42
  "stage_names": [
@@ -59,12 +59,13 @@
59
  "window_block_indexes": [
60
  0,
61
  1,
62
- 3,
63
  4,
64
- 6,
65
  7,
66
- 9,
67
- 10
 
68
  ]
69
  },
70
  "bbox_cost": 5,
@@ -72,13 +73,13 @@
72
  "c2f_num_blocks": 3,
73
  "class_cost": 2,
74
  "class_loss_coefficient": 1,
75
- "d_model": 384,
76
  "decoder_activation_function": "relu",
77
- "decoder_cross_attention_heads": 24,
78
  "decoder_ffn_dim": 2048,
79
- "decoder_layers": 3,
80
- "decoder_n_points": 4,
81
- "decoder_self_attention_heads": 12,
82
  "dice_loss_coefficient": 1,
83
  "disable_custom_kernels": true,
84
  "dropout": 0.1,
@@ -283,17 +284,15 @@
283
  "mask_loss_coefficient": 1,
284
  "mask_point_sample_ratio": 16,
285
  "model_type": "rf_detr",
286
- "num_feature_levels": 2,
287
  "num_queries": 300,
288
  "projector_in_channels": [
289
- 384,
290
- 384
291
  ],
292
- "projector_out_channels": 384,
293
  "projector_scale_factors": [
294
- 2.0,
295
- 0.5
296
  ],
297
  "segmentation_head_activation_function": "gelu",
298
- "transformers_version": "5.0.0.dev0"
299
  }
 
13
  "drop_path_rate": 0.0,
14
  "hidden_act": "gelu",
15
  "hidden_dropout_prob": 0.0,
16
+ "hidden_size": 384,
17
+ "image_size": 704,
18
  "initializer_range": 0.02,
19
  "layer_norm_eps": 1e-06,
20
  "layerscale_value": 1.0,
21
  "mlp_ratio": 4,
22
  "model_type": "rf_detr_dinov2",
23
+ "num_attention_heads": 6,
24
  "num_channels": 3,
25
  "num_hidden_layers": 12,
26
+ "num_windows": 2,
27
  "out_features": [
28
+ "stage3",
29
+ "stage6",
30
+ "stage9",
31
+ "stage12"
32
  ],
33
  "out_indices": [
34
+ 3,
35
+ 6,
36
+ 9,
37
+ 12
38
  ],
39
+ "patch_size": 16,
40
  "qkv_bias": true,
41
  "reshape_hidden_states": true,
42
  "stage_names": [
 
59
  "window_block_indexes": [
60
  0,
61
  1,
62
+ 2,
63
  4,
64
+ 5,
65
  7,
66
+ 8,
67
+ 10,
68
+ 11
69
  ]
70
  },
71
  "bbox_cost": 5,
 
73
  "c2f_num_blocks": 3,
74
  "class_cost": 2,
75
  "class_loss_coefficient": 1,
76
+ "d_model": 256,
77
  "decoder_activation_function": "relu",
78
+ "decoder_cross_attention_heads": 16,
79
  "decoder_ffn_dim": 2048,
80
+ "decoder_layers": 4,
81
+ "decoder_n_points": 2,
82
+ "decoder_self_attention_heads": 8,
83
  "dice_loss_coefficient": 1,
84
  "disable_custom_kernels": true,
85
  "dropout": 0.1,
 
284
  "mask_loss_coefficient": 1,
285
  "mask_point_sample_ratio": 16,
286
  "model_type": "rf_detr",
287
+ "num_feature_levels": 1,
288
  "num_queries": 300,
289
  "projector_in_channels": [
290
+ 256
 
291
  ],
292
+ "projector_out_channels": 256,
293
  "projector_scale_factors": [
294
+ 1.0
 
295
  ],
296
  "segmentation_head_activation_function": "gelu",
297
+ "transformers_version": "5.0.1.dev0"
298
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b7ad3cc3c67e41679d0630f2c98095551d4a287ffe796c135c21dbea6ba84a2
3
- size 542617560
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:366c6e28361817aad7fc4da97db2ce400db177bea7a2f5eb9f7d324e5a07d46f
3
+ size 135796200