iliasslasri commited on
Commit
535d8f0
·
verified ·
1 Parent(s): 673617f

Chess Challenge submission by iliasslasri

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. config.json +6 -6
  3. model.safetensors +2 -2
README.md CHANGED
@@ -14,13 +14,13 @@ Chess model submitted to the LLM Course Chess Challenge.
14
  ## Submission Info
15
 
16
  - **Submitted by**: [iliasslasri](https://huggingface.co/iliasslasri)
17
- - **Parameters**: 997,136
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
  - **Vocab size**: 75
24
- - **Embedding dim**: 96
25
  - **Layers**: 11
26
- - **Heads**: 8
 
14
  ## Submission Info
15
 
16
  - **Submitted by**: [iliasslasri](https://huggingface.co/iliasslasri)
17
+ - **Parameters**: 980,720
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
  - **Vocab size**: 75
24
+ - **Embedding dim**: 92
25
  - **Layers**: 11
26
+ - **Heads**: 4
config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "_name_or_path": "./gqa_1_ft/checkpoint-719934/",
3
  "architectures": [
4
  "ChessForCausalLM"
5
  ],
6
- "attn": "GQA",
7
  "auto_map": {
8
  "AutoConfig": "model.ChessConfig",
9
  "AutoModelForCausalLM": "model.ChessForCausalLM"
@@ -14,11 +14,11 @@
14
  "layer_norm_epsilon": 1e-05,
15
  "model_type": "chess_transformer",
16
  "n_ctx": 256,
17
- "n_embd": 96,
18
- "n_head": 8,
19
- "n_inner": 304,
20
  "n_layer": 11,
21
- "num_groups": 4,
22
  "pad_token_id": 0,
23
  "tie_weights": false,
24
  "tie_word_embeddings": false,
 
1
  {
2
+ "_name_or_path": "./11_4_92_ft_ft_ft/checkpoint-475008/",
3
  "architectures": [
4
  "ChessForCausalLM"
5
  ],
6
+ "attn": "MHA",
7
  "auto_map": {
8
  "AutoConfig": "model.ChessConfig",
9
  "AutoModelForCausalLM": "model.ChessForCausalLM"
 
14
  "layer_norm_epsilon": 1e-05,
15
  "model_type": "chess_transformer",
16
  "n_ctx": 256,
17
+ "n_embd": 92,
18
+ "n_head": 4,
19
+ "n_inner": 276,
20
  "n_layer": 11,
21
+ "num_groups": 2,
22
  "pad_token_id": 0,
23
  "tie_weights": false,
24
  "tie_word_embeddings": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0b7e13ee34b41230e39c989ce65f47c313da964b2d1fecb0c27bd1feccde1890
3
- size 4003888
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f7244a5c854e9c9684f98b1b63970ad82899c0545f1fb1b105ce1ae2e8f76a8
3
+ size 3934384