xingjianleng commited on
Commit
fa0cf31
·
verified ·
1 Parent(s): eaed631

Delete f16d64/args.json

Browse files
Files changed (1) hide show
  1. f16d64/args.json +0 -61
f16d64/args.json DELETED
@@ -1,61 +0,0 @@
1
- {
2
- "output_dir": "exps",
3
- "exp_name": "sit-xl-1-invae-f16d64-dinov2-vit-b-enc8-coeff0.5-cosine-mlp-400k",
4
- "logging_dir": "logs",
5
- "report_to": "wandb",
6
- "sampling_steps": 50000,
7
- "resume_step": 0,
8
- "n_samples": 64,
9
- "sync_checkpoints": false,
10
- "model": "SiT-XL/1",
11
- "num_classes": 1000,
12
- "encoder_depth": 8,
13
- "fused_attn": true,
14
- "qk_norm": false,
15
- "projection_layer_type": "mlp",
16
- "proj_kwargs_kernel_size": 1,
17
- "vae": "f16d64",
18
- "vae_ckpt": "pretrained_models/invae-f16d64.pt",
19
- "vae_latents_name": "invae-f16d64",
20
- "data_dir": "/scratch/xingjian.leng/data",
21
- "imagenet_dir": "data",
22
- "resolution": 256,
23
- "batch_size": 256,
24
- "vis_enc_resolution": 256,
25
- "precached_latents_dir": null,
26
- "allow_tf32": true,
27
- "mixed_precision": "fp16",
28
- "epochs": 1400,
29
- "max_train_steps": 400000,
30
- "checkpointing_steps": 25000,
31
- "gradient_accumulation_steps": 1,
32
- "learning_rate": 0.0001,
33
- "adam_beta1": 0.9,
34
- "adam_beta2": 0.999,
35
- "adam_weight_decay": 0.0,
36
- "adam_epsilon": 1e-08,
37
- "max_grad_norm": 1.0,
38
- "compile": false,
39
- "ema_decay": 0.9999,
40
- "ema_update_freq": 1,
41
- "seed": 0,
42
- "num_workers": 16,
43
- "path_type": "linear",
44
- "prediction": "v",
45
- "cfg_prob": 0.1,
46
- "enc_type": "dinov2-vit-b",
47
- "proj_coeff": "0.5",
48
- "weighting": "uniform",
49
- "repa_loss": true,
50
- "projection_loss_type": "cosine",
51
- "spnorm_method": "none",
52
- "cls_token_weight": 0.2,
53
- "zscore_alpha": 1.0,
54
- "zscore_proj_skip_std": false,
55
- "symmetric": true,
56
- "tau": 0.5,
57
- "negatives": "intra",
58
- "bef_proj": false,
59
- "remove_diag": true,
60
- "instancenorm": false
61
- }