Upload runs/ub5iiz6d/config.json with huggingface_hub
Browse files- runs/ub5iiz6d/config.json +37 -0
runs/ub5iiz6d/config.json
ADDED
|
@@ -0,0 +1,37 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"dataset": {
|
| 3 |
+
"max_k": 5,
|
| 4 |
+
"max_seq_len": 12,
|
| 5 |
+
"samples_per_k": 200000,
|
| 6 |
+
"task": "s5",
|
| 7 |
+
"test_size": 0.2
|
| 8 |
+
},
|
| 9 |
+
"model": {
|
| 10 |
+
"d_model": 128,
|
| 11 |
+
"dropout": 0.1,
|
| 12 |
+
"max_seq_len": 12,
|
| 13 |
+
"nhead": 4,
|
| 14 |
+
"num_classes": 120,
|
| 15 |
+
"num_layers": 4,
|
| 16 |
+
"num_tokens": 123,
|
| 17 |
+
"type": "GroupTransformer",
|
| 18 |
+
"use_compile": true
|
| 19 |
+
},
|
| 20 |
+
"task": "s5",
|
| 21 |
+
"train": {
|
| 22 |
+
"batch_size": 2048,
|
| 23 |
+
"beta1": 0.9,
|
| 24 |
+
"beta2": 0.999,
|
| 25 |
+
"gradient_clip": 1.0,
|
| 26 |
+
"lr": "1e-4",
|
| 27 |
+
"max_epochs_per_stage": 200,
|
| 28 |
+
"max_val_acc": 0.99,
|
| 29 |
+
"op_eps": "1e-8",
|
| 30 |
+
"weight_decay": 0.01
|
| 31 |
+
},
|
| 32 |
+
"wandb": {
|
| 33 |
+
"name": "S5-GroupTransformer-L4-H4-D128-seq12-k5-s200000-bs2048-lr1e-4-wd0.01",
|
| 34 |
+
"run_id": "ub5iiz6d",
|
| 35 |
+
"url": "https://wandb.ai/bkitano/nc1-tc0-transformer-toy-experiments/runs/ub5iiz6d"
|
| 36 |
+
}
|
| 37 |
+
}
|