eyaa99 commited on
Commit
d861bb9
·
verified ·
1 Parent(s): 5d36c92

Chess Challenge submission by eyaa99

Browse files
Files changed (4) hide show
  1. README.md +3 -3
  2. config.json +3 -3
  3. model.safetensors +2 -2
  4. vocab.json +0 -0
README.md CHANGED
@@ -14,13 +14,13 @@ Chess model submitted to the LLM Course Chess Challenge.
14
  ## Submission Info
15
 
16
  - **Submitted by**: [eyaa99](https://huggingface.co/eyaa99)
17
- - **Parameters**: 909,824
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
- - **Vocab size**: 1682
24
  - **Embedding dim**: 128
25
- - **Layers**: 4
26
  - **Heads**: 4
 
14
  ## Submission Info
15
 
16
  - **Submitted by**: [eyaa99](https://huggingface.co/eyaa99)
17
+ - **Parameters**: 947,968
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
+ - **Vocab size**: 3272
24
  - **Embedding dim**: 128
25
+ - **Layers**: 3
26
  - **Heads**: 4
config.json CHANGED
@@ -3,7 +3,7 @@
3
  "ChessForCausalLM"
4
  ],
5
  "bos_token_id": 1,
6
- "dropout": 0.1,
7
  "dtype": "float32",
8
  "eos_token_id": 2,
9
  "layer_norm_epsilon": 1e-05,
@@ -12,9 +12,9 @@
12
  "n_embd": 128,
13
  "n_head": 4,
14
  "n_inner": 384,
15
- "n_layer": 4,
16
  "pad_token_id": 0,
17
  "tie_weights": true,
18
  "transformers_version": "4.57.6",
19
- "vocab_size": 1682
20
  }
 
3
  "ChessForCausalLM"
4
  ],
5
  "bos_token_id": 1,
6
+ "dropout": 0.05,
7
  "dtype": "float32",
8
  "eos_token_id": 2,
9
  "layer_norm_epsilon": 1e-05,
 
12
  "n_embd": 128,
13
  "n_head": 4,
14
  "n_inner": 384,
15
+ "n_layer": 3,
16
  "pad_token_id": 0,
17
  "tie_weights": true,
18
  "transformers_version": "4.57.6",
19
+ "vocab_size": 3272
20
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4dfa72d7a980612daddba42a056272ba950629ea312c59a45926a2e28d287b83
3
- size 3643696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8634bc794af37f7764c6d7b02706371c5d02056bd03dfdaad1ebf060664656cb
3
+ size 3795248
vocab.json CHANGED
The diff for this file is too large to render. See raw diff