--- library_name: transformers tags: - gpt2 - text-generation --- # Model Card for harpertoken/harpertokenGPT2 GPT-2 small model trained from scratch on WikiText-2-raw-v1 dataset for text generation. ## Model Details ### Model Description This is a GPT-2 small model (117M parameters) trained from random initialization on the WikiText-2-raw-v1 dataset. It can generate coherent text continuations. - **Developed by:** Niladri Das - **Model type:** GPT-2 - **Language(s) (NLP):** English - **License:** Apache-2.0 ### Model Sources - **Repository:** https://github.com/bniladridas/models ## Uses ### Direct Use Use for text generation tasks, such as completing sentences or generating stories. ### Out-of-Scope Use Not suitable for tasks requiring factual accuracy, safety-critical applications, or languages other than English. ## Bias, Risks, and Limitations Trained on WikiText, which may contain biases from the source data. Model may generate inappropriate or biased content. ### Recommendations Use with caution; implement content filters for production use. ## How to Get Started with the Model ```python from transformers import pipeline generator = pipeline('text-generation', model='harpertoken/harpertokenGPT2') print(generator("The quick brown fox")) ``` ## Training Details ### Training Data WikiText-2-raw-v1 dataset, a collection of Wikipedia articles. ### Training Procedure Trained from scratch using PyTorch and Transformers. #### Training Hyperparameters - Epochs: 3 - Batch size: 1 - Learning rate: 5e-5 - Max length: 512 ## Evaluation Basic evaluation via text generation coherence. ### Results Generates plausible text continuations. ## Environmental Impact - **Hardware Type:** CPU/MPS - **Hours used:** ~10 minutes - **Carbon Emitted:** Minimal (local training) ## Technical Specifications ### Model Architecture and Objective GPT-2 decoder-only transformer for causal language modeling. ### Compute Infrastructure - Hardware: Mac with MPS - Software: PyTorch, Transformers