|
|
--- |
|
|
tags: |
|
|
- text-generation |
|
|
- gpt2 |
|
|
license: gpl-3.0 |
|
|
datasets: |
|
|
- bookcorpus |
|
|
metrics: |
|
|
- perplexity |
|
|
model-index: |
|
|
- name: basic-text-generator |
|
|
results: |
|
|
- task: |
|
|
type: text-generation |
|
|
dataset: |
|
|
name: bookcorpus |
|
|
type: bookcorpus |
|
|
metrics: |
|
|
- name: Perplexity |
|
|
type: perplexity |
|
|
value: 25.3 |
|
|
--- |
|
|
|
|
|
# Basic Text Generator |
|
|
|
|
|
## Overview |
|
|
|
|
|
This is a fine-tuned GPT-2 model for general text generation. It can continue prompts, generate stories, or create coherent paragraphs based on input text. Trained on a diverse corpus for broad applicability. |
|
|
|
|
|
## Model Architecture |
|
|
|
|
|
- Base Model: GPT-2 |
|
|
- Layers: 12 |
|
|
- Hidden Size: 768 |
|
|
- Attention Heads: 12 |
|
|
- Context Window: 1024 tokens |
|
|
|
|
|
## Intended Use |
|
|
|
|
|
Suitable for creative writing, content generation, or prototyping language-based applications. |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- May generate biased or inappropriate content based on training data. |
|
|
- Outputs can be repetitive or nonsensical for long generations. |
|
|
- Not optimized for specific domains like code or math. |
|
|
|
|
|
## Example Code |
|
|
|
|
|
```python |
|
|
from transformers import pipeline |
|
|
|
|
|
generator = pipeline("text-generation", model="user/basic-text-generator") |
|
|
result = generator("Once upon a time,", max_length=50) |
|
|
print(result[0]['generated_text']) |
|
|
# "Once upon a time, in a land far away..." |