Spaces:
Build error
Build error
clip-path
Browse files
configs/image_cond_diffuser_asl/image-ASLDM-256.yaml
CHANGED
|
@@ -24,7 +24,7 @@ model:
|
|
| 24 |
target: michelangelo.models.tsal.clip_asl_module.CLIPAlignedShapeAsLatentModule
|
| 25 |
params:
|
| 26 |
# clip_model_version: "./checkpoints/clip/clip-vit-large-patch14"
|
| 27 |
-
clip_model_version: "
|
| 28 |
|
| 29 |
loss_cfg:
|
| 30 |
target: torch.nn.Identity
|
|
@@ -33,7 +33,7 @@ model:
|
|
| 33 |
target: michelangelo.models.conditional_encoders.encoder_factory.FrozenCLIPImageGridEmbedder
|
| 34 |
params:
|
| 35 |
# version: "./checkpoints/clip/clip-vit-large-patch14"
|
| 36 |
-
version: "
|
| 37 |
zero_embedding_radio: 0.1
|
| 38 |
|
| 39 |
first_stage_key: "surface"
|
|
|
|
| 24 |
target: michelangelo.models.tsal.clip_asl_module.CLIPAlignedShapeAsLatentModule
|
| 25 |
params:
|
| 26 |
# clip_model_version: "./checkpoints/clip/clip-vit-large-patch14"
|
| 27 |
+
clip_model_version: "checkpoints/clip/clip-vit-large-patch14"
|
| 28 |
|
| 29 |
loss_cfg:
|
| 30 |
target: torch.nn.Identity
|
|
|
|
| 33 |
target: michelangelo.models.conditional_encoders.encoder_factory.FrozenCLIPImageGridEmbedder
|
| 34 |
params:
|
| 35 |
# version: "./checkpoints/clip/clip-vit-large-patch14"
|
| 36 |
+
version: "checkpoints/clip/clip-vit-large-patch14"
|
| 37 |
zero_embedding_radio: 0.1
|
| 38 |
|
| 39 |
first_stage_key: "surface"
|
configs/text_cond_diffuser_asl/text-ASLDM-256.yaml
CHANGED
|
@@ -23,7 +23,7 @@ model:
|
|
| 23 |
aligned_module_cfg:
|
| 24 |
target: michelangelo.models.tsal.clip_asl_module.CLIPAlignedShapeAsLatentModule
|
| 25 |
params:
|
| 26 |
-
clip_model_version: "
|
| 27 |
|
| 28 |
loss_cfg:
|
| 29 |
target: torch.nn.Identity
|
|
@@ -31,7 +31,7 @@ model:
|
|
| 31 |
cond_stage_config:
|
| 32 |
target: michelangelo.models.conditional_encoders.encoder_factory.FrozenAlignedCLIPTextEmbedder
|
| 33 |
params:
|
| 34 |
-
version: "
|
| 35 |
zero_embedding_radio: 0.1
|
| 36 |
max_length: 77
|
| 37 |
|
|
|
|
| 23 |
aligned_module_cfg:
|
| 24 |
target: michelangelo.models.tsal.clip_asl_module.CLIPAlignedShapeAsLatentModule
|
| 25 |
params:
|
| 26 |
+
clip_model_version: "checkpoints/clip/clip-vit-large-patch14"
|
| 27 |
|
| 28 |
loss_cfg:
|
| 29 |
target: torch.nn.Identity
|
|
|
|
| 31 |
cond_stage_config:
|
| 32 |
target: michelangelo.models.conditional_encoders.encoder_factory.FrozenAlignedCLIPTextEmbedder
|
| 33 |
params:
|
| 34 |
+
version: "checkpoints/clip/clip-vit-large-patch14"
|
| 35 |
zero_embedding_radio: 0.1
|
| 36 |
max_length: 77
|
| 37 |
|