File size: 774 Bytes
eee46f1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
{
  "architecture": "Raschka GPTModel (separate W_query/W_key/W_value, no weight tying)",
  "model_type": "instruction-tuned (SFT)",
  "base_model": "nishantup/nanogpt-slm-124m (gpt_slm_best.pth)",
  "model_config": {
    "vocab_size": 50257,
    "context_length": 256,
    "emb_dim": 768,
    "n_heads": 12,
    "n_layers": 12,
    "drop_rate": 0.0,
    "qkv_bias": false
  },
  "total_parameters_millions": 163.2,
  "tokenizer": "tiktoken gpt2 (50,257 BPE tokens)",
  "framework": "PyTorch",
  "prompt_format": "Alpaca (### Instruction / ### Input / ### Response)",
  "training": {
    "dataset": "Alpaca-format instruction dataset (1,100 examples)",
    "epochs": 2,
    "optimizer": "AdamW (lr=5e-5, weight_decay=0.1)",
    "max_length": 256
  }
}