focusondepth / config.json
Younes Belkada
add model
d6de736
raw
history blame contribute delete
552 Bytes
{
"_name_or_path": "ybelkada/focusondepth",
"architectures": [
"FocusOnDepth"
],
"emb_dim": 768,
"hooks": [
2,
5,
8,
11
],
"image_size": [
3,
384,
384
],
"model_timm": "vit_base_patch16_384",
"model_type": "focusondepth",
"nclasses": 2,
"num_layers_encoder": 24,
"patch_size": 16,
"read": "projection",
"reassemble_s": [
4,
8,
16,
32
],
"resample_dim": 256,
"torch_dtype": "float32",
"transformer_dropout": 0,
"transformers_version": "4.16.2",
"type_": "full"
}