dev-das commited on
Commit
afb5828
·
verified ·
1 Parent(s): ef3097a

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +9 -11
config.json CHANGED
@@ -1,17 +1,15 @@
1
  {
2
- "architectures": [
3
- "MyGPTForCausalLM"
4
- ],
5
- "context_length": 256,
6
- "drop_rate": 0.1,
7
- "dtype": "float32",
8
- "emb_dim": 256,
9
  "model_type": "my_gpt",
10
- "n_heads": 4,
11
- "n_layers": 12,
12
- "qkv_bias": false,
13
- "transformers_version": "5.1.0",
14
  "vocab_size": 50257,
 
 
 
 
 
 
 
15
 
16
  "bos_token_id": 50256,
17
  "eos_token_id": 50256,
 
1
  {
2
+ "architectures": ["MyGPTForCausalLM"],
 
 
 
 
 
 
3
  "model_type": "my_gpt",
4
+
 
 
 
5
  "vocab_size": 50257,
6
+ "hidden_size": 256,
7
+ "num_attention_heads": 4,
8
+ "num_hidden_layers": 12,
9
+ "max_position_embeddings": 256,
10
+
11
+ "drop_rate": 0.1,
12
+ "qkv_bias": false,
13
 
14
  "bos_token_id": 50256,
15
  "eos_token_id": 50256,