iliasslasri commited on
Commit
4448949
·
verified ·
1 Parent(s): 36dc8ee

Chess Challenge submission by iliasslasri

Browse files
Files changed (3) hide show
  1. README.md +3 -3
  2. config.json +6 -6
  3. model.safetensors +2 -2
README.md CHANGED
@@ -14,13 +14,13 @@ Chess model submitted to the LLM Course Chess Challenge.
14
  ## Submission Info
15
 
16
  - **Submitted by**: [iliasslasri](https://huggingface.co/iliasslasri)
17
- - **Parameters**: 980,720
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
  - **Vocab size**: 75
24
- - **Embedding dim**: 92
25
  - **Layers**: 11
26
- - **Heads**: 4
 
14
  ## Submission Info
15
 
16
  - **Submitted by**: [iliasslasri](https://huggingface.co/iliasslasri)
17
+ - **Parameters**: 997,136
18
  - **Organization**: LLM-course
19
 
20
  ## Model Details
21
 
22
  - **Architecture**: Chess Transformer (GPT-style)
23
  - **Vocab size**: 75
24
+ - **Embedding dim**: 96
25
  - **Layers**: 11
26
+ - **Heads**: 8
config.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "_name_or_path": "./11_4_92_ft_ft_ft/checkpoint-411921/",
3
  "architectures": [
4
  "ChessForCausalLM"
5
  ],
6
- "attn": "MHA",
7
  "auto_map": {
8
  "AutoConfig": "model.ChessConfig",
9
  "AutoModelForCausalLM": "model.ChessForCausalLM"
@@ -14,11 +14,11 @@
14
  "layer_norm_epsilon": 1e-05,
15
  "model_type": "chess_transformer",
16
  "n_ctx": 256,
17
- "n_embd": 92,
18
- "n_head": 4,
19
- "n_inner": 276,
20
  "n_layer": 11,
21
- "num_groups": 2,
22
  "pad_token_id": 0,
23
  "tie_weights": false,
24
  "tie_word_embeddings": false,
 
1
  {
2
+ "_name_or_path": "./gqa_1/checkpoint-341412/",
3
  "architectures": [
4
  "ChessForCausalLM"
5
  ],
6
+ "attn": "GQA",
7
  "auto_map": {
8
  "AutoConfig": "model.ChessConfig",
9
  "AutoModelForCausalLM": "model.ChessForCausalLM"
 
14
  "layer_norm_epsilon": 1e-05,
15
  "model_type": "chess_transformer",
16
  "n_ctx": 256,
17
+ "n_embd": 96,
18
+ "n_head": 8,
19
+ "n_inner": 304,
20
  "n_layer": 11,
21
+ "num_groups": 4,
22
  "pad_token_id": 0,
23
  "tie_weights": false,
24
  "tie_word_embeddings": false,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:acf4e3c08c84faca194f0b7f0b9053b222458bdf7a7c516fc49d225aa7ae4a33
3
- size 3934384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f569dc1198e833f8e1923bcf79f61d5b50dc6d343527f363fe4f210e3ff45ab5
3
+ size 4003888