ZeroCool94 commited on
Commit
ff953e9
·
1 Parent(s): deede35

Delete maskgit.4090000.pt.yaml

Browse files
Files changed (1) hide show
  1. maskgit.4090000.pt.yaml +0 -74
maskgit.4090000.pt.yaml DELETED
@@ -1,74 +0,0 @@
1
- only_save_last_checkpoint: false
2
- validation_image_scale: 1.0
3
- no_center_crop: false
4
- no_flip: false
5
- dataset_save_path: E:\cached_datasets\INE2
6
- clear_previous_experiments: false
7
- num_tokens: 8192
8
- seq_len: 1024
9
- depth: 4
10
- dim_head: 64
11
- heads: 8
12
- ff_mult: 4
13
- t5_name: t5-large
14
- cond_image_size: null
15
- validation_prompt: a girl|a cat|a dog
16
- timesteps: 18
17
- max_grad_norm: null
18
- seed: 42
19
- valid_frac: 0.05
20
- use_ema: false
21
- ema_beta: 0.995
22
- ema_update_after_step: 1
23
- ema_update_every: 1
24
- apply_grad_penalty_every: 4
25
- image_column: image
26
- caption_column: caption
27
- log_with: wandb
28
- mixed_precision: 'no'
29
- use_8bit_adam: false
30
- results_dir: results\Muse-v0.9
31
- logging_dir: null
32
- vae_path: results\Muse-v0.9\vae.9420000.pt
33
- dataset_name: null
34
- hf_split_name: null
35
- streaming: false
36
- train_data_dir: E:\dataset
37
- num_train_steps: -1
38
- num_epochs: 100
39
- dim: 32
40
- batch_size: 1
41
- lr: 1.0e-05
42
- gradient_accumulation_steps: 100
43
- save_results_every: 500
44
- save_model_every: 5000
45
- checkpoint_limit: null
46
- vq_codebook_size: 8192
47
- vq_codebook_dim: 8192
48
- cond_drop_prob: 0.5
49
- image_size: 256
50
- lr_scheduler: constant_with_warmup
51
- scheduler_power: 1.0
52
- lr_warmup_steps: 1
53
- num_cycles: 1
54
- resume_path: results\Muse-v0.9\maskgit.4050000.pt
55
- taming_model_path: null
56
- taming_config_path: null
57
- optimizer: Adam
58
- weight_decay: 0.045
59
- cache_path: null
60
- no_cache: true
61
- link: false
62
- latest_checkpoint: true
63
- do_not_save_config: false
64
- use_l2_recon_loss: false
65
- debug: false
66
- config_path: null
67
- attention_type: ein
68
- random_crop: true
69
- project_name: muse_maskgit
70
- run_name: null
71
- wandb_user: sygil
72
- channels: 3
73
- layers: 4
74
- discr_layers: 4