Kiy-K commited on
Commit
83027d5
·
verified ·
1 Parent(s): 30afdcd

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +47 -0
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language: en
3
+ license: mit
4
+ library_name: transformers
5
+ tags:
6
+ - chess
7
+ - mamba
8
+ - moe
9
+ ---
10
+
11
+ # ♟️ KiyEngine V3 (Mamba-MoE)
12
+
13
+ Chess engine powered by Mamba-MoE architecture. Final training loss: **5.46**
14
+
15
+ ## Model Architecture
16
+
17
+ - **Type**: Mamba with Mixture of Experts (MoE)
18
+ - **Parameters**:
19
+ - d_model: 384
20
+ - n_layers: 4
21
+ - n_experts: 8
22
+ - top_k: 2
23
+ - d_state: 16
24
+ - d_conv: 4
25
+ - expansion_factor: 2
26
+ - vocab_size: 768
27
+
28
+ ## Usage
29
+
30
+ ```python
31
+ from transformers import AutoConfig, AutoModel
32
+
33
+ # Load the model
34
+ config = AutoConfig.from_pretrained("Kiy-K/KiyEngine-V3-Mamba-MoE", trust_remote_code=True)
35
+ model = AutoModel.from_pretrained("Kiy-K/KiyEngine-V3-Mamba-MoE", trust_remote_code=True)
36
+ ```
37
+
38
+ ## Training Details
39
+
40
+ Model trained on chess data with final loss of 5.46.
41
+
42
+ ## Files
43
+
44
+ - `model.safetensors`: Model weights
45
+ - `config.json`: Model configuration
46
+ - `configuration_kiyengine.py`: Configuration class definition
47
+ - `modeling_kiyengine.py`: Model implementation (if available)