File size: 387 Bytes
c803403
 
 
 
 
 
 
 
 
 
d0b63db
 
 
 
 
 
 
c803403
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
{
  "num_opinion_anchors": 225,
  "pentachoron_dim": 512,
  "scales": [
    128,
    256,
    512
  ],
  "vocab_size": 492,
  "max_seq_len": 77,
  "global_step": 1000,
  "best_val_loss": Infinity,
  "optimizations": {
    "use_gradient_checkpointing": false,
    "share_scale_embeddings": false
  },
  "note": "share_scale_embeddings MUST be False to preserve multi-scale architecture"
}