Upload prior/prior_config.json with huggingface_hub
Browse files- prior/prior_config.json +10 -9
prior/prior_config.json
CHANGED
|
@@ -8,7 +8,7 @@
|
|
| 8 |
"dim": 768,
|
| 9 |
"depth": 12,
|
| 10 |
"num_timesteps": 1000,
|
| 11 |
-
"max_text_len":
|
| 12 |
"num_time_embeds": 1,
|
| 13 |
"num_image_embeds": 1,
|
| 14 |
"num_text_embeds": 1,
|
|
@@ -34,15 +34,15 @@
|
|
| 34 |
"condition_on_text_encodings": true
|
| 35 |
},
|
| 36 |
"data": {
|
| 37 |
-
"batch_size":
|
| 38 |
"num_data_points": 50805086,
|
| 39 |
"eval_every_seconds": 1800,
|
| 40 |
"image_url": "s3://s-datasets/laion-aesthetic/embeddings/img_emb",
|
| 41 |
"meta_url": "s3://s-datasets/laion-aesthetic/embeddings/metadata",
|
| 42 |
"splits": {
|
| 43 |
"train": 0.9,
|
| 44 |
-
"val":
|
| 45 |
-
"test":0.
|
| 46 |
}
|
| 47 |
},
|
| 48 |
"train": {
|
|
@@ -52,13 +52,14 @@
|
|
| 52 |
"max_grad_norm": 0.5,
|
| 53 |
"use_ema": true,
|
| 54 |
"ema_beta": 0.9999,
|
| 55 |
-
"ema_update_after_step":
|
| 56 |
"amp": false,
|
| 57 |
"save_every_seconds": 3600,
|
| 58 |
-
"eval_timesteps": [64, 1000]
|
|
|
|
| 59 |
},
|
| 60 |
"tracker": {
|
| 61 |
-
"data_path": ".prior-updates",
|
| 62 |
"overwrite_data_path": true,
|
| 63 |
"log": {
|
| 64 |
"log_type": "wandb",
|
|
@@ -74,8 +75,8 @@
|
|
| 74 |
{
|
| 75 |
"save_to": "local",
|
| 76 |
"save_type": "checkpoint",
|
| 77 |
-
"save_latest_to": ".prior-updates/latest_checkpoint.pth",
|
| 78 |
-
"save_best_to": ".prior-updates/best_checkpoint.pth"
|
| 79 |
},
|
| 80 |
{
|
| 81 |
"save_to": "huggingface",
|
|
|
|
| 8 |
"dim": 768,
|
| 9 |
"depth": 12,
|
| 10 |
"num_timesteps": 1000,
|
| 11 |
+
"max_text_len": 77,
|
| 12 |
"num_time_embeds": 1,
|
| 13 |
"num_image_embeds": 1,
|
| 14 |
"num_text_embeds": 1,
|
|
|
|
| 34 |
"condition_on_text_encodings": true
|
| 35 |
},
|
| 36 |
"data": {
|
| 37 |
+
"batch_size": 215,
|
| 38 |
"num_data_points": 50805086,
|
| 39 |
"eval_every_seconds": 1800,
|
| 40 |
"image_url": "s3://s-datasets/laion-aesthetic/embeddings/img_emb",
|
| 41 |
"meta_url": "s3://s-datasets/laion-aesthetic/embeddings/metadata",
|
| 42 |
"splits": {
|
| 43 |
"train": 0.9,
|
| 44 |
+
"val": 9.6e-5,
|
| 45 |
+
"test":0.09990399999999998
|
| 46 |
}
|
| 47 |
},
|
| 48 |
"train": {
|
|
|
|
| 52 |
"max_grad_norm": 0.5,
|
| 53 |
"use_ema": true,
|
| 54 |
"ema_beta": 0.9999,
|
| 55 |
+
"ema_update_after_step": 20,
|
| 56 |
"amp": false,
|
| 57 |
"save_every_seconds": 3600,
|
| 58 |
+
"eval_timesteps": [64, 1000],
|
| 59 |
+
"random_seed": 84513
|
| 60 |
},
|
| 61 |
"tracker": {
|
| 62 |
+
"data_path": ".prior-updates-dropout",
|
| 63 |
"overwrite_data_path": true,
|
| 64 |
"log": {
|
| 65 |
"log_type": "wandb",
|
|
|
|
| 75 |
{
|
| 76 |
"save_to": "local",
|
| 77 |
"save_type": "checkpoint",
|
| 78 |
+
"save_latest_to": ".prior-updates-dropout/latest_checkpoint.pth",
|
| 79 |
+
"save_best_to": ".prior-updates-dropout/best_checkpoint.pth"
|
| 80 |
},
|
| 81 |
{
|
| 82 |
"save_to": "huggingface",
|