Text Generation
Transformers
PyTorch
Safetensors
English
gpt2
NoSleep
Reddit
Story
Horror
text-generation-inference
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("DarwinAnim8or/NoSleepPromptGen")
model = AutoModelForCausalLM.from_pretrained("DarwinAnim8or/NoSleepPromptGen")Quick Links
"NoSleep" Writing Prompt Generator
Finetuned version of GPT2 to facilitate generation of Writing Prompts for the GPT-NoSleep-355m model
You can use the space linked on the right to use this model, then use the NoSleep model in tandem to generate stories!
Training Procedure
This was trained on the 'reddt-nosleep-posts' dataset, using the "HappyTransformers" library on Google Colab. This model was trained for X epochs with learning rate 1e-2.
Biases & Limitations
This likely contains the same biases and limitations as the original GPT2 that it is based on, and additionally heavy biases from the dataset. It likely will generate offensive output.
Intended Use
This model is meant for fun, nothing else.
Sample Use
from happytransformer import GENSettings
args_top_k = GENSettings(no_repeat_ngram_size=1, do_sample=True, top_k=80, temperature=0.4, max_length=25, early_stopping=True)
result = happy_gen.generate_text("[WP] \"", args=args_top_k)
print(result.text)
- Downloads last month
- 7
# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="DarwinAnim8or/NoSleepPromptGen")