Chess Challenge submission by gbl1357
Browse files- README.md +3 -3
- config.json +3 -3
- model.safetensors +2 -2
README.md
CHANGED
|
@@ -14,13 +14,13 @@ Chess model submitted to the LLM Course Chess Challenge.
|
|
| 14 |
## Submission Info
|
| 15 |
|
| 16 |
- **Submitted by**: [gbl1357](https://huggingface.co/gbl1357)
|
| 17 |
-
- **Parameters**:
|
| 18 |
- **Organization**: LLM-course
|
| 19 |
|
| 20 |
## Model Details
|
| 21 |
|
| 22 |
- **Architecture**: Chess Transformer (GPT-style)
|
| 23 |
- **Vocab size**: 73
|
| 24 |
-
- **Embedding dim**:
|
| 25 |
-
- **Layers**:
|
| 26 |
- **Heads**: 4
|
|
|
|
| 14 |
## Submission Info
|
| 15 |
|
| 16 |
- **Submitted by**: [gbl1357](https://huggingface.co/gbl1357)
|
| 17 |
+
- **Parameters**: 950,872
|
| 18 |
- **Organization**: LLM-course
|
| 19 |
|
| 20 |
## Model Details
|
| 21 |
|
| 22 |
- **Architecture**: Chess Transformer (GPT-style)
|
| 23 |
- **Vocab size**: 73
|
| 24 |
+
- **Embedding dim**: 104
|
| 25 |
+
- **Layers**: 7
|
| 26 |
- **Heads**: 4
|
config.json
CHANGED
|
@@ -9,10 +9,10 @@
|
|
| 9 |
"layer_norm_epsilon": 1e-05,
|
| 10 |
"model_type": "chess_transformer",
|
| 11 |
"n_ctx": 256,
|
| 12 |
-
"n_embd":
|
| 13 |
"n_head": 4,
|
| 14 |
-
"n_inner":
|
| 15 |
-
"n_layer":
|
| 16 |
"pad_token_id": 0,
|
| 17 |
"tie_weights": true,
|
| 18 |
"transformers_version": "4.57.5",
|
|
|
|
| 9 |
"layer_norm_epsilon": 1e-05,
|
| 10 |
"model_type": "chess_transformer",
|
| 11 |
"n_ctx": 256,
|
| 12 |
+
"n_embd": 104,
|
| 13 |
"n_head": 4,
|
| 14 |
+
"n_inner": 416,
|
| 15 |
+
"n_layer": 7,
|
| 16 |
"pad_token_id": 0,
|
| 17 |
"tie_weights": true,
|
| 18 |
"transformers_version": "4.57.5",
|
model.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0ad5282db5804df9e7274693b000313b9288e26c581a3ac31c64c9a57cb195b4
|
| 3 |
+
size 3809784
|