DedeProGames commited on
Commit
ed4e972
·
verified ·
1 Parent(s): 5adc36a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +110 -54
README.md CHANGED
@@ -1,54 +1,110 @@
1
- ---
2
- library_name: transformers
3
- license: mit
4
- base_model: microsoft/DialoGPT-small
5
- tags:
6
- - generated_from_trainer
7
- model-index:
8
- - name: dialochess
9
- results: []
10
- ---
11
-
12
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
- should probably proofread and complete it, then remove this comment. -->
14
-
15
- # dialochess
16
-
17
- This model is a fine-tuned version of [microsoft/DialoGPT-small](https://huggingface.co/microsoft/DialoGPT-small) on an unknown dataset.
18
-
19
- ## Model description
20
-
21
- More information needed
22
-
23
- ## Intended uses & limitations
24
-
25
- More information needed
26
-
27
- ## Training and evaluation data
28
-
29
- More information needed
30
-
31
- ## Training procedure
32
-
33
- ### Training hyperparameters
34
-
35
- The following hyperparameters were used during training:
36
- - learning_rate: 5e-05
37
- - train_batch_size: 8
38
- - eval_batch_size: 8
39
- - seed: 42
40
- - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
41
- - lr_scheduler_type: cosine
42
- - num_epochs: 1
43
- - mixed_precision_training: Native AMP
44
-
45
- ### Training results
46
-
47
-
48
-
49
- ### Framework versions
50
-
51
- - Transformers 4.52.4
52
- - Pytorch 2.7.1+cu118
53
- - Datasets 3.6.0
54
- - Tokenizers 0.21.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: mit
4
+ base_model: microsoft/DialoGPT-small
5
+ tags:
6
+ - generated_from_trainer
7
+ - dialog
8
+ - chess
9
+ - gpt-2
10
+ model-index:
11
+ - name: dialochess
12
+ results: []
13
+ ---
14
+ # Model card: **dialochess**
15
+
16
+ **Short description**
17
+ `dialochess` is a conversational model derived from `microsoft/DialoGPT-small` and fine-tuned for chess-related dialog and play. It was trained using the dataset referenced by the trainer (see *Training data* below) and configured to interact in chess play/analysis settings such as the Hugging Face Space `mlabonne/chessllm`.
18
+
19
+ ---
20
+
21
+ ## Model details
22
+
23
+ - **Model type:** GPT-2 / DialoGPT-small family (causal, autoregressive)
24
+ - **Base model:** `microsoft/DialoGPT-small`
25
+ - **Fine-tuned name:** `dialochess`
26
+ - **License:** MIT
27
+ - **Libraries & versions used during training (reported by trainer):**
28
+ - Transformers 4.52.4
29
+ - PyTorch 2.7.1+cu118
30
+ - Datasets 3.6.0
31
+ - Tokenizers 0.21.1
32
+
33
+ ---
34
+
35
+ ## Model description
36
+
37
+ `dialochess` is a fine-tuned conversational transformer intended to produce chess-relevant dialogue: move suggestions, commentary, brief explanations of positions, and to play short games against other AIs. The model remains an autoregressive language model (not a dedicated chess engine) and generates text tokens that may include algebraic moves, evaluation phrases, and natural-language commentary.
38
+
39
+ ---
40
+
41
+ ## Intended uses & limitations
42
+
43
+ **Intended uses**
44
+ - Research and experimentation in conversational chess agents.
45
+ - Integration into chat-based chess interfaces for move suggestions and commentary.
46
+ - Generating sample game dialogues or annotated move lists for educational/demo purposes.
47
+ - Fine-tuning baseline for further chess-specific language-model work.
48
+
49
+ **Not suitable for**
50
+ - Replacing a dedicated chess engine for precise tactical calculation (e.g., Stockfish, Leela).
51
+ - High-stakes or competitive play where rigorous move correctness and deep search are required.
52
+ - Any medical, legal, financial, or safety-critical advice — it's a domain-specific conversational model and may hallucinate or produce incorrect information.
53
+
54
+ **Limitations**
55
+ - May hallucinate moves, annotations, or claims about positions.
56
+ - Performance is dependent on the quality and diversity of the fine-tuning dataset (see *Training and evaluation data*).
57
+ - No official evaluation metrics were included in the automatically-generated card. Users should validate with specific benchmarks (perplexity, move-accuracy, Elo/win-rate against baselines).
58
+
59
+ ---
60
+
61
+ ## Training and evaluation data
62
+
63
+ - **Dataset (provided by the trainer):** The fine-tuning dataset referenced in the trainer materials and available in the provided Colab notebook:
64
+ `https://colab.research.google.com/drive/11UjbfajCzphe707_V7PD-2e5WIzyintf`
65
+ (This link was included by the trainer. Please review the Colab to inspect dataset sources, composition, license, and any preprocessing steps.)
66
+ - **Source / provenance:** The model was trained to interact with or play against other AIs in the Hugging Face Space `mlabonne/chessllm`. See the space here: `https://huggingface.co/spaces/mlabonne/chessllm`.
67
+ - **Data filtering & cleaning:** Not provided in the auto-generated metadata. It is recommended to include details about tokenization choices, any filtering of illegal moves or metadata removal, and train/validation splits.
68
+ - **Privacy & licenses:** The original trainer metadata did not list dataset license(s). Verify that any third-party game logs, PGN files, or scraped content used for training are permitted under their licenses before public redistribution.
69
+
70
+ ---
71
+
72
+ ## Training procedure
73
+
74
+ **Hyperparameters (as reported):**
75
+ - learning_rate: 5e-05
76
+ - train_batch_size: 8
77
+ - eval_batch_size: 8
78
+ - seed: 42
79
+ - optimizer: AdamW (betas=(0.9, 0.999), eps=1e-08)
80
+ - lr_scheduler_type: cosine
81
+ - num_epochs: 1
82
+ - mixed_precision_training: Native AMP
83
+
84
+ **Notes**
85
+ - The model was fine-tuned from `microsoft/DialoGPT-small`. The training ran for 1 epoch (per provided metadata). For improved performance, consider longer training, larger datasets, and careful evaluation/early stopping.
86
+ - No detailed training logs or evaluation metrics were included in the auto-generated card; add validation loss curves, perplexity, and any chess-specific metrics (move prediction accuracy, legality rate, win-rate vs baseline) to the card if available.
87
+
88
+ ---
89
+
90
+ ## Evaluation & recommended metrics
91
+
92
+ No evaluation results were included in the auto-generated card. To assess quality, we recommend reporting:
93
+ - **Perplexity** on a held-out validation set.
94
+ - **Move accuracy**: fraction of model-predicted moves that match the recorded moves in a test PGN corpus.
95
+ - **Legal-move rate**: fraction of generated moves that are legal given the position.
96
+ - **Win-rate / Elo proxy**: Play matches against a fixed baseline agent and report win/draw/loss and Elo-like estimates.
97
+ - **Human preference / qualitative eval**: human raters judge helpfulness, fluency, and chess correctness in dialog samples.
98
+
99
+ If you want, run a small evaluation pipeline and paste the results here so this section can be updated.
100
+
101
+ ---
102
+
103
+ ## Safety & biases
104
+
105
+ - The model can generate incorrect or misleading chess content. Verify generated moves with a chess engine before acting on them.
106
+ - As an autoregressive language model, it may reproduce biases or toxic language present in the training data. Use standard moderation / filtering if deploying publicly.
107
+ - Avoid exposing the model as a canonical authority on chess positions or instructing users to rely on it without verification.
108
+
109
+ ---
110
+