--- license: cc0-1.0 datasets: - Navanjana/Gutenberg_books - aisuko/simple_english_wikipedia - stas/openwebtext-10k - RaiBP/openwebtext2-first-30-chunks-lang-detect-raw-output - lucadiliello/bookcorpusopen - deepmind/pg19 language: - en pipeline_tag: text-generation library_name: transformers tags: - Self model-index: - name: AaI results: - task: type: text-classification name: Multiple Choice dataset: name: ai2_arc type: ai2_arc config: ARC-Easy split: test metrics: - name: Accuracy type: accuracy value: 0.8 --- ## **Safety Concerns** This model has not passed any safety tuning. We are not responsible for any damages. We updated this model from .pth to .safetensors. ## AaI Introduction AaI is a model fully made by 16dvnk on his NVIDIA Geforce RTX 4080 Laptop GPU. He trained it for 11 hours straight, and after some tuning, has made this model. The model is made from scratch. He claims the process was a pain, and has taken lots of effort. He named it AaI and not AAI or other variations since he thinks it is an “eyesore”. ## Architecture The model uses a Generative pre-trained transformer architecture. ## Technical Specifications | AaI Specs | Details | |------------------------|----------------------------------------| | Creator | 16dvnk | | Hardware | NVIDIA GeForce RTX 4080 Laptop GPU | | Training Duration | 11 hours | | Framework | PyTorch | | Parameter Count | 14 million | | Model Type | Generative pre-trained transformer | | Initial Training Year | 2025 | | Stable Release Status | No stable release as of September 2025| ## Evaluation Results The model was evaluated on the **ARC-Easy** benchmark (test split). | Dataset | Split | Metric | Value | |----------|-------|----------|---------| | ARC-Easy | test | Accuracy | 0.80% | ## Notes • All current releases have 14M parameters, which is considered small. • The model was trained using PyTorch. • As of September 2025, there is no stable release of AaI.