{ "_attn_implementation_autoset": true, "architectures": [ "FIMSDE" ], "dropout_rate": 0.1, "dtype": "float32", "finetune": false, "finetune_detach_diffusion": false, "finetune_em_steps": 1, "finetune_num_points": -1, "finetune_on_sampling_mse": false, "finetune_on_sampling_nll": false, "finetune_samples_count": 1, "finetune_samples_steps": 1, "hidden_act": { "name": "torch.nn.GELU" }, "hidden_layers": [ 256, 256 ], "learnable_loss_scales": { "attention": { "activation": "torch.nn.GELU", "dim_feedforward": 1024, "dropout": 0.1, "nhead": 8 }, "num_res_layers": 8, "paths_block_attention": false, "projection": { "dropout": 0.1, "hidden_act": { "name": "torch.nn.GELU" }, "hidden_layers": [ 256, 256 ], "name": "fim.models.blocks.base.MLP" } }, "max_dimension": 3, "model_embedding_size": 256, "model_type": "fimsde", "name": "FIMSDE", "num_locations_on_path": 32, "operator": { "attention": { "activation": "torch.nn.GELU", "dim_feedforward": 1024, "dropout": 0.1, "nhead": 8 }, "num_res_layers": 8, "paths_block_attention": false, "projection": { "dropout": 0.1, "hidden_act": { "name": "torch.nn.GELU" }, "hidden_layers": [ 256, 256 ], "name": "fim.models.blocks.base.MLP" } }, "phi_0t": { "name": "torch.nn.Linear" }, "phi_0x": { "name": "torch.nn.Linear" }, "phi_1x": { "name": "torch.nn.Linear" }, "psi_1": { "layer": { "activation": "torch.nn.GELU", "attn_method": "linear", "dim_feedforward": 1024, "dropout": 0.1, "lin_feature_map": "softmax", "lin_normalize": false, "nhead": 8 }, "name": "CombinedPathTransformer", "num_layers": 2 }, "residual_ff_size": 1024, "states_norm": { "name": "fim.models.sde.Standardization" }, "times_norm": { "name": "fim.models.sde.DeltaLogCentering" }, "transformer_layer_activation": "gelu", "transformers_version": "4.57.1" }