--- library_name: transformers tags: - chess - llm-course - chess-challenge license: mit --- # chess-ooooooooo A chess transformer model trained for the LLM Course Chess Challenge. ## Model Architecture This model uses a GPT-style transformer architecture optimized for chess move prediction: - **Parameters**: 948,352 (0.95M) - **Vocabulary size**: 85 - **Embedding dimension**: 128 - **Number of layers**: 6 - **Attention heads**: 4 - **Feed-forward dimension**: 320 - **Context length**: 256 - **Dropout**: 0.1 ## Training The model was trained on a subset of the Lichess 2025 dataset, focusing on learning valid chess move sequences. The architecture was carefully tuned to stay within the 1M parameter constraint while maintaining reasonable performance. ## Usage ```python from transformers import AutoModelForCausalLM from src.tokenizer import ChessTokenizer model = AutoModelForCausalLM.from_pretrained( "LLM-course/chess-ooooooooo", trust_remote_code=True ) tokenizer = ChessTokenizer.from_pretrained( "LLM-course/chess-ooooooooo", trust_remote_code=True ) # Generate moves input_text = "[BOS] WPe2e4" input_ids = tokenizer.encode(input_text) outputs = model.generate(input_ids, max_length=50) predicted_moves = tokenizer.decode(outputs[0]) ``` ## Submission Submitted by [etienneLefranc](https://huggingface.co/etienneLefranc) for the LLM Course Chess Challenge.