Configuration Parsing
Warning:
Invalid JSON for config file config.json
YAML Metadata
Warning:
empty or missing yaml metadata in repo card
(https://huggingface.co/docs/hub/model-cards#model-card-metadata)
TinyTextGenerator
Overview
TinyTextGenerator is a small causal language model based on the GPT-2 architecture, designed for basic text generation tasks. With only 6 layers, it is lightweight and fast, making it ideal for experimentation, local deployment, or educational use.
Model Architecture
- Model type: GPT-2 (causal language modeling)
- Hidden size: 768
- Number of layers: 6
- Number of attention heads: 12
- Vocabulary size: 50,257
- Context length: 1024 tokens
- Parameters: ~82M
Built using GPT2LMHeadModel from the Transformers library.
Usage
from transformers import pipeline
generator = pipeline(
"text-generation",
model="your-username/TinyTextGenerator"
)
output = generator(
"The future of AI is",
max_new_tokens=50,
do_sample=True,
top_p=0.95
)
print(output[0]['generated_text'])
- Downloads last month
- 22
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support