{ "activation": "gelu", "arch": "trans_enc", "condition": "text", "dropout": 0.1, "ff_size": 1024, "flip_sin_to_cos": true, "freq_shift": 0, "guidance_scale": 7.5, "guidance_uncondp": 0.1, "latent_dim": [ 1, 256 ], "nclasses": 10, "nfeats": 263, "normalize_before": false, "num_heads": 4, "num_layers": 9, "pe_type": "mld", "position_embedding": "learned", "return_intermediate_dec": false, "skip_connect": true, "text_encoded_dim": 768 }