| { |
| "architectures": [ |
| "HieraModel" |
| ], |
| "dim_mul": 2.0, |
| "drop_path_rate": 0.0, |
| "embedding_dimension": 96, |
| "head_dropout": 0.0, |
| "head_init_scale": 0.001, |
| "head_mul": 2.0, |
| "in_chans": 3, |
| "input_size": [ |
| 224, |
| 224 |
| ], |
| "mask_unit_attn": [ |
| true, |
| true, |
| false, |
| false |
| ], |
| "mask_unit_size": [ |
| 8, |
| 8 |
| ], |
| "mlp_ratio": 4.0, |
| "model_type": "hiera", |
| "num_classes": 1000, |
| "number_of_heads": 1, |
| "patch_kernel": [ |
| 7, |
| 7 |
| ], |
| "patch_padding": [ |
| 3, |
| 3 |
| ], |
| "patch_stride": [ |
| 4, |
| 4 |
| ], |
| "q_pool": 3, |
| "q_stride": [ |
| 2, |
| 2 |
| ], |
| "sep_position_embeddings": false, |
| "stages": [ |
| 2, |
| 3, |
| 16, |
| 3 |
| ], |
| "torch_dtype": "float32", |
| "transformers_version": "4.39.0.dev0" |
| } |
|
|