Text Generation
Transformers
Safetensors
Spanish
English
llama
text-generation-inference
unsloth
trl
sft
conversational
# Load model directly
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("krory/GenBook-Deepseek-R1.Llama-8B")
model = AutoModelForCausalLM.from_pretrained("krory/GenBook-Deepseek-R1.Llama-8B")
messages = [
{"role": "user", "content": "Who are you?"},
]
inputs = tokenizer.apply_chat_template(
messages,
add_generation_prompt=True,
tokenize=True,
return_dict=True,
return_tensors="pt",
).to(model.device)
outputs = model.generate(**inputs, max_new_tokens=40)
print(tokenizer.decode(outputs[0][inputs["input_ids"].shape[-1]:]))Quick Links
About the Model
This model is designed to be a storytelling AI capable of creating fun, engaging, and well-structured narratives. Its purpose is to serve as an interactive tool for generating and experiencing unique stories in real time, tailored to the user's input and preferences.
Key Features
- Interactive Narratives: Produces coherent and entertaining stories based on user prompts, adapting dynamically to maintain engagement.
- Consistent World-Building: Ensures logical progression and consistency in characters, settings, and events across long narratives.
- Optimized for Efficiency: Built to perform reliably on limited hardware while delivering high-quality outputs.
Training Overview
The model was fine-tuned using datasets focused on narrative construction, character development, and immersive descriptions. Key aspects of the training include:
- Adaptability: Special attention was given to creating a system that responds flexibly to varied user inputs while maintaining coherence.
- Resource Efficiency: Techniques like LoRA (Low-Rank Adaptation) and 4-bit quantization were employed to optimize memory usage without compromising output quality.
- Long-Context Support: Enhanced with methods to handle extended interactions and complex storylines.
Purpose
The primary goal of this model is to create a personal, customizable storytelling AI, allowing users to immerse themselves in unique, AI-driven stories anytime.
- Downloads last month
- 6

# Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-generation", model="krory/GenBook-Deepseek-R1.Llama-8B") messages = [ {"role": "user", "content": "Who are you?"}, ] pipe(messages)