Sunxt25 commited on
Commit
0f30676
·
verified ·
1 Parent(s): 5592107

Chess Challenge submission by Sunxt25

Browse files
Files changed (5) hide show
  1. README.md +6 -6
  2. config.json +6 -9
  3. model.safetensors +2 -2
  4. tokenizer_config.json +33 -8
  5. vocab.json +6 -1
README.md CHANGED
@@ -7,20 +7,20 @@ tags:
7
  license: mit
8
  ---
9
 
10
- # chess-Sunxt25
11
 
12
  Chess model submitted to the LLM Course Chess Challenge.
13
 
14
  ## Submission Info
15
 
16
  - **Submitted by**: [Sunxt25](https://huggingface.co/Sunxt25)
17
- - **Parameters**: 878,336
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
- - **Vocab size**: 144
24
- - **Embedding dim**: 128
25
- - **Layers**: 5
26
- - **Heads**: 4
 
7
  license: mit
8
  ---
9
 
10
+ # chess-sunxt25
11
 
12
  Chess model submitted to the LLM Course Chess Challenge.
13
 
14
  ## Submission Info
15
 
16
  - **Submitted by**: [Sunxt25](https://huggingface.co/Sunxt25)
17
+ - **Parameters**: 921,480
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
+ - **Vocab size**: 149
24
+ - **Embedding dim**: 120
25
+ - **Layers**: 6
26
+ - **Heads**: 8
config.json CHANGED
@@ -2,9 +2,6 @@
2
  "architectures": [
3
  "ChessForCausalLM"
4
  ],
5
- "auto_map": {
6
- "AutoTokenizer": "chess_tokenizer_custom.ChessTokenizer"
7
- },
8
  "bos_token_id": 1,
9
  "dropout": 0.1,
10
  "dtype": "float32",
@@ -12,12 +9,12 @@
12
  "layer_norm_epsilon": 1e-05,
13
  "model_type": "chess_transformer",
14
  "n_ctx": 256,
15
- "n_embd": 128,
16
- "n_head": 4,
17
- "n_inner": 384,
18
- "n_layer": 5,
19
  "pad_token_id": 0,
20
  "tie_weights": true,
21
  "transformers_version": "4.57.5",
22
- "vocab_size": 144
23
- }
 
2
  "architectures": [
3
  "ChessForCausalLM"
4
  ],
 
 
 
5
  "bos_token_id": 1,
6
  "dropout": 0.1,
7
  "dtype": "float32",
 
9
  "layer_norm_epsilon": 1e-05,
10
  "model_type": "chess_transformer",
11
  "n_ctx": 256,
12
+ "n_embd": 120,
13
+ "n_head": 8,
14
+ "n_inner": 360,
15
+ "n_layer": 6,
16
  "pad_token_id": 0,
17
  "tie_weights": true,
18
  "transformers_version": "4.57.5",
19
+ "vocab_size": 149
20
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2b168eb150974e5eff71ecdc4c7dc06a708917e31fc17001362a4550da9da07
3
- size 3518768
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:010cb99e29355984a756f59f47dd9bda7c4e3945f0f6600d54eb6bff6ec75ef1
3
+ size 3692360
tokenizer_config.json CHANGED
@@ -1,12 +1,37 @@
1
  {
2
  "added_tokens_decoder": {
3
- "0": { "content": "[PAD]", "special": true },
4
- "1": { "content": "[BOS]", "special": true },
5
- "2": { "content": "[EOS]", "special": true },
6
- "3": { "content": "[UNK]", "special": true }
7
- },
8
- "auto_map": {
9
- "AutoTokenizer": "chess_tokenizer_custom.ChessTokenizer"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
10
  },
11
  "bos_token": "[BOS]",
12
  "clean_up_tokenization_spaces": false,
@@ -16,4 +41,4 @@
16
  "pad_token": "[PAD]",
17
  "tokenizer_class": "ChessTokenizer",
18
  "unk_token": "[UNK]"
19
- }
 
1
  {
2
  "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[BOS]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[EOS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ }
35
  },
36
  "bos_token": "[BOS]",
37
  "clean_up_tokenization_spaces": false,
 
41
  "pad_token": "[PAD]",
42
  "tokenizer_class": "ChessTokenizer",
43
  "unk_token": "[UNK]"
44
+ }
vocab.json CHANGED
@@ -142,5 +142,10 @@
142
  "e8_t": 140,
143
  "f8_t": 141,
144
  "g8_t": 142,
145
- "h8_t": 143
 
 
 
 
 
146
  }
 
142
  "e8_t": 140,
143
  "f8_t": 141,
144
  "g8_t": 142,
145
+ "h8_t": 143,
146
+ "(x)": 144,
147
+ "(+)": 145,
148
+ "(+*)": 146,
149
+ "(o)": 147,
150
+ "(O)": 148
151
  }