Add config from convert_d_fine_original_pytorch_checkpoint_to_hf.py
Browse files- config.json +9 -5
config.json
CHANGED
|
@@ -1,9 +1,10 @@
|
|
| 1 |
{
|
|
|
|
| 2 |
"activation_dropout": 0.0,
|
| 3 |
"activation_function": "silu",
|
| 4 |
-
"anchor_image_size":
|
| 5 |
-
|
| 6 |
-
|
| 7 |
],
|
| 8 |
"attention_dropout": 0.0,
|
| 9 |
"auxiliary_loss": true,
|
|
@@ -73,7 +74,11 @@
|
|
| 73 |
],
|
| 74 |
"decoder_layers": 4,
|
| 75 |
"decoder_method": "default",
|
| 76 |
-
"decoder_n_points":
|
|
|
|
|
|
|
|
|
|
|
|
|
| 77 |
"decoder_offset_scale": 0.5,
|
| 78 |
"depth_mult": 0.67,
|
| 79 |
"disable_custom_kernels": true,
|
|
@@ -290,7 +295,6 @@
|
|
| 290 |
"positional_encoding_temperature": 10000,
|
| 291 |
"reg_scale": 4.0,
|
| 292 |
"top_prob_values": 4,
|
| 293 |
-
"torch_dtype": "float32",
|
| 294 |
"transformers_version": "4.50.0.dev0",
|
| 295 |
"use_focal_loss": true,
|
| 296 |
"use_pretrained_backbone": false,
|
|
|
|
| 1 |
{
|
| 2 |
+
"_attn_implementation_autoset": true,
|
| 3 |
"activation_dropout": 0.0,
|
| 4 |
"activation_function": "silu",
|
| 5 |
+
"anchor_image_size": [
|
| 6 |
+
640,
|
| 7 |
+
640
|
| 8 |
],
|
| 9 |
"attention_dropout": 0.0,
|
| 10 |
"auxiliary_loss": true,
|
|
|
|
| 74 |
],
|
| 75 |
"decoder_layers": 4,
|
| 76 |
"decoder_method": "default",
|
| 77 |
+
"decoder_n_points": [
|
| 78 |
+
3,
|
| 79 |
+
6,
|
| 80 |
+
3
|
| 81 |
+
],
|
| 82 |
"decoder_offset_scale": 0.5,
|
| 83 |
"depth_mult": 0.67,
|
| 84 |
"disable_custom_kernels": true,
|
|
|
|
| 295 |
"positional_encoding_temperature": 10000,
|
| 296 |
"reg_scale": 4.0,
|
| 297 |
"top_prob_values": 4,
|
|
|
|
| 298 |
"transformers_version": "4.50.0.dev0",
|
| 299 |
"use_focal_loss": true,
|
| 300 |
"use_pretrained_backbone": false,
|