Add rf-detr-large model
Browse files- config.json +28 -29
- model.safetensors +2 -2
config.json
CHANGED
|
@@ -13,30 +13,30 @@
|
|
| 13 |
"drop_path_rate": 0.0,
|
| 14 |
"hidden_act": "gelu",
|
| 15 |
"hidden_dropout_prob": 0.0,
|
| 16 |
-
"hidden_size":
|
| 17 |
-
"image_size":
|
| 18 |
"initializer_range": 0.02,
|
| 19 |
"layer_norm_eps": 1e-06,
|
| 20 |
"layerscale_value": 1.0,
|
| 21 |
"mlp_ratio": 4,
|
| 22 |
"model_type": "rf_detr_dinov2",
|
| 23 |
-
"num_attention_heads":
|
| 24 |
"num_channels": 3,
|
| 25 |
"num_hidden_layers": 12,
|
| 26 |
-
"num_windows":
|
| 27 |
"out_features": [
|
| 28 |
-
"
|
| 29 |
-
"
|
| 30 |
-
"
|
| 31 |
-
"
|
| 32 |
],
|
| 33 |
"out_indices": [
|
| 34 |
-
|
| 35 |
-
|
| 36 |
-
|
| 37 |
-
|
| 38 |
],
|
| 39 |
-
"patch_size":
|
| 40 |
"qkv_bias": true,
|
| 41 |
"reshape_hidden_states": true,
|
| 42 |
"stage_names": [
|
|
@@ -59,12 +59,13 @@
|
|
| 59 |
"window_block_indexes": [
|
| 60 |
0,
|
| 61 |
1,
|
| 62 |
-
|
| 63 |
4,
|
| 64 |
-
|
| 65 |
7,
|
| 66 |
-
|
| 67 |
-
10
|
|
|
|
| 68 |
]
|
| 69 |
},
|
| 70 |
"bbox_cost": 5,
|
|
@@ -72,13 +73,13 @@
|
|
| 72 |
"c2f_num_blocks": 3,
|
| 73 |
"class_cost": 2,
|
| 74 |
"class_loss_coefficient": 1,
|
| 75 |
-
"d_model":
|
| 76 |
"decoder_activation_function": "relu",
|
| 77 |
-
"decoder_cross_attention_heads":
|
| 78 |
"decoder_ffn_dim": 2048,
|
| 79 |
-
"decoder_layers":
|
| 80 |
-
"decoder_n_points":
|
| 81 |
-
"decoder_self_attention_heads":
|
| 82 |
"dice_loss_coefficient": 1,
|
| 83 |
"disable_custom_kernels": true,
|
| 84 |
"dropout": 0.1,
|
|
@@ -283,17 +284,15 @@
|
|
| 283 |
"mask_loss_coefficient": 1,
|
| 284 |
"mask_point_sample_ratio": 16,
|
| 285 |
"model_type": "rf_detr",
|
| 286 |
-
"num_feature_levels":
|
| 287 |
"num_queries": 300,
|
| 288 |
"projector_in_channels": [
|
| 289 |
-
|
| 290 |
-
384
|
| 291 |
],
|
| 292 |
-
"projector_out_channels":
|
| 293 |
"projector_scale_factors": [
|
| 294 |
-
|
| 295 |
-
0.5
|
| 296 |
],
|
| 297 |
"segmentation_head_activation_function": "gelu",
|
| 298 |
-
"transformers_version": "5.0.
|
| 299 |
}
|
|
|
|
| 13 |
"drop_path_rate": 0.0,
|
| 14 |
"hidden_act": "gelu",
|
| 15 |
"hidden_dropout_prob": 0.0,
|
| 16 |
+
"hidden_size": 384,
|
| 17 |
+
"image_size": 704,
|
| 18 |
"initializer_range": 0.02,
|
| 19 |
"layer_norm_eps": 1e-06,
|
| 20 |
"layerscale_value": 1.0,
|
| 21 |
"mlp_ratio": 4,
|
| 22 |
"model_type": "rf_detr_dinov2",
|
| 23 |
+
"num_attention_heads": 6,
|
| 24 |
"num_channels": 3,
|
| 25 |
"num_hidden_layers": 12,
|
| 26 |
+
"num_windows": 2,
|
| 27 |
"out_features": [
|
| 28 |
+
"stage3",
|
| 29 |
+
"stage6",
|
| 30 |
+
"stage9",
|
| 31 |
+
"stage12"
|
| 32 |
],
|
| 33 |
"out_indices": [
|
| 34 |
+
3,
|
| 35 |
+
6,
|
| 36 |
+
9,
|
| 37 |
+
12
|
| 38 |
],
|
| 39 |
+
"patch_size": 16,
|
| 40 |
"qkv_bias": true,
|
| 41 |
"reshape_hidden_states": true,
|
| 42 |
"stage_names": [
|
|
|
|
| 59 |
"window_block_indexes": [
|
| 60 |
0,
|
| 61 |
1,
|
| 62 |
+
2,
|
| 63 |
4,
|
| 64 |
+
5,
|
| 65 |
7,
|
| 66 |
+
8,
|
| 67 |
+
10,
|
| 68 |
+
11
|
| 69 |
]
|
| 70 |
},
|
| 71 |
"bbox_cost": 5,
|
|
|
|
| 73 |
"c2f_num_blocks": 3,
|
| 74 |
"class_cost": 2,
|
| 75 |
"class_loss_coefficient": 1,
|
| 76 |
+
"d_model": 256,
|
| 77 |
"decoder_activation_function": "relu",
|
| 78 |
+
"decoder_cross_attention_heads": 16,
|
| 79 |
"decoder_ffn_dim": 2048,
|
| 80 |
+
"decoder_layers": 4,
|
| 81 |
+
"decoder_n_points": 2,
|
| 82 |
+
"decoder_self_attention_heads": 8,
|
| 83 |
"dice_loss_coefficient": 1,
|
| 84 |
"disable_custom_kernels": true,
|
| 85 |
"dropout": 0.1,
|
|
|
|
| 284 |
"mask_loss_coefficient": 1,
|
| 285 |
"mask_point_sample_ratio": 16,
|
| 286 |
"model_type": "rf_detr",
|
| 287 |
+
"num_feature_levels": 1,
|
| 288 |
"num_queries": 300,
|
| 289 |
"projector_in_channels": [
|
| 290 |
+
256
|
|
|
|
| 291 |
],
|
| 292 |
+
"projector_out_channels": 256,
|
| 293 |
"projector_scale_factors": [
|
| 294 |
+
1.0
|
|
|
|
| 295 |
],
|
| 296 |
"segmentation_head_activation_function": "gelu",
|
| 297 |
+
"transformers_version": "5.0.1.dev0"
|
| 298 |
}
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:366c6e28361817aad7fc4da97db2ce400db177bea7a2f5eb9f7d324e5a07d46f
|
| 3 |
+
size 135796200
|