vladislavbro commited on
Commit
9e29218
·
verified ·
1 Parent(s): 1091987

Add config from convert_d_fine_original_pytorch_checkpoint_to_hf.py

Browse files
Files changed (1) hide show
  1. config.json +34 -10
config.json CHANGED
@@ -1,10 +1,8 @@
1
  {
 
2
  "activation_dropout": 0.0,
3
  "activation_function": "silu",
4
  "anchor_image_size": null,
5
- "architectures": [
6
- "DFineForObjectDetection"
7
- ],
8
  "attention_dropout": 0.0,
9
  "auxiliary_loss": true,
10
  "backbone": null,
@@ -21,6 +19,29 @@
21
  2,
22
  3,
23
  4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  ]
25
  },
26
  "backbone_kwargs": null,
@@ -31,13 +52,17 @@
31
  "decoder_attention_heads": 8,
32
  "decoder_ffn_dim": 1024,
33
  "decoder_in_channels": [
34
- 256,
35
- 256,
36
- 256
37
  ],
38
  "decoder_layers": 6,
39
  "decoder_method": "default",
40
- "decoder_n_points": 4,
 
 
 
 
41
  "decoder_offset_scale": 0.5,
42
  "depth_mult": 1.0,
43
  "disable_custom_kernels": true,
@@ -47,8 +72,8 @@
47
  ],
48
  "encoder_activation_function": "gelu",
49
  "encoder_attention_heads": 8,
50
- "encoder_ffn_dim": 1024,
51
- "encoder_hidden_dim": 256,
52
  "encoder_in_channels": [
53
  512,
54
  1024,
@@ -254,7 +279,6 @@
254
  "positional_encoding_temperature": 10000,
255
  "reg_scale": 4.0,
256
  "top_prob_values": 4,
257
- "torch_dtype": "float32",
258
  "transformers_version": "4.50.0.dev0",
259
  "use_focal_loss": true,
260
  "use_pretrained_backbone": false,
 
1
  {
2
+ "_attn_implementation_autoset": true,
3
  "activation_dropout": 0.0,
4
  "activation_function": "silu",
5
  "anchor_image_size": null,
 
 
 
6
  "attention_dropout": 0.0,
7
  "auxiliary_loss": true,
8
  "backbone": null,
 
19
  2,
20
  3,
21
  4
22
+ ],
23
+ "stage_in_channels": [
24
+ 64,
25
+ 128,
26
+ 512,
27
+ 1024
28
+ ],
29
+ "stage_mid_channels": [
30
+ 64,
31
+ 128,
32
+ 256,
33
+ 512
34
+ ],
35
+ "stage_num_blocks": [
36
+ 1,
37
+ 2,
38
+ 5,
39
+ 2
40
+ ],
41
+ "stem_channels": [
42
+ 3,
43
+ 32,
44
+ 64
45
  ]
46
  },
47
  "backbone_kwargs": null,
 
52
  "decoder_attention_heads": 8,
53
  "decoder_ffn_dim": 1024,
54
  "decoder_in_channels": [
55
+ 384,
56
+ 384,
57
+ 384
58
  ],
59
  "decoder_layers": 6,
60
  "decoder_method": "default",
61
+ "decoder_n_points": [
62
+ 3,
63
+ 6,
64
+ 3
65
+ ],
66
  "decoder_offset_scale": 0.5,
67
  "depth_mult": 1.0,
68
  "disable_custom_kernels": true,
 
72
  ],
73
  "encoder_activation_function": "gelu",
74
  "encoder_attention_heads": 8,
75
+ "encoder_ffn_dim": 2048,
76
+ "encoder_hidden_dim": 384,
77
  "encoder_in_channels": [
78
  512,
79
  1024,
 
279
  "positional_encoding_temperature": 10000,
280
  "reg_scale": 4.0,
281
  "top_prob_values": 4,
 
282
  "transformers_version": "4.50.0.dev0",
283
  "use_focal_loss": true,
284
  "use_pretrained_backbone": false,