|
|
--- |
|
|
language: |
|
|
- en |
|
|
tags: |
|
|
- autotrain |
|
|
- text-generation |
|
|
- causal-lm |
|
|
- lora |
|
|
- style-transfer |
|
|
- agatha-christie |
|
|
license: other |
|
|
datasets: |
|
|
- realdanielbyrne/AgathaChristieText |
|
|
base_model: google/gemma-2-2b-it |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
--- |
|
|
|
|
|
# 🕵️ Agatha Christie Style LoRA — Gemma-2-2B-IT |
|
|
|
|
|
This model is a **LoRA fine-tuned** version of **Google’s Gemma-2-2B-IT**, trained to generate prose in the **classic style of Agatha Christie**. |
|
|
It captures the tone, pacing, domestic atmosphere, and subtle psychological tension characteristic of Christie’s detective fiction. |
|
|
|
|
|
The model is intended for **story continuation**, **creative writing**, and **mystery scene generation**. |
|
|
It is a **causal LM**, not an instruction-tuned chat model. |
|
|
|
|
|
--- |
|
|
|
|
|
## ✨ Model Details |
|
|
|
|
|
- **Model type:** Causal language model (LoRA adapter) |
|
|
- **Base model:** `google/gemma-2-2b-it` |
|
|
- **Fine-tuning method:** LoRA (PEFT) |
|
|
- **Training framework:** Hugging Face AutoTrain |
|
|
- **Languages:** English |
|
|
- **Primary domain:** Fiction, mystery, detective prose |
|
|
- **Task:** Text generation (story continuation) |
|
|
|
|
|
--- |
|
|
|
|
|
## 🎯 Intended Use |
|
|
|
|
|
This model is suitable for: |
|
|
|
|
|
- Generating text in the Agatha Christie style |
|
|
- Continuing mystery scenes |
|
|
- Creative writing assistance |
|
|
- Prototyping detective stories |
|
|
- Experimenting with stylistic transfer |
|
|
|
|
|
### Not recommended for: |
|
|
|
|
|
- Real-world Q&A |
|
|
- Safety-critical applications |
|
|
- Factual reasoning |
|
|
- Chat-assistant behavior without custom prompting |
|
|
|
|
|
--- |
|
|
|
|
|
## 📚 Training Data |
|
|
|
|
|
The model was trained on: |
|
|
|
|
|
- **`realdanielbyrne/AgathaChristieText`** |
|
|
A plain-text dataset containing ~13,000 rows of Agatha Christie prose. |
|
|
|
|
|
No instruction template was used; the model was trained purely on **unstructured text** to learn style rather than task formats. |
|
|
|
|
|
--- |
|
|
|
|
|
## 🔧 Training Configuration |
|
|
|
|
|
- **Epochs:** 3 |
|
|
- **Block size:** 1024 |
|
|
- **Learning rate:** 3e-5 |
|
|
- **Batch size:** AutoTrain default |
|
|
- **Optimizer:** AdamW |
|
|
- **Chat template:** *None (causal LM)* |
|
|
- **LoRA target modules:** Linear projection layers of Gemma |
|
|
|
|
|
Training was performed using **AutoTrain**, which handled preprocessing, batching, and evaluation. |
|
|
|
|
|
--- |
|
|
|
|
|
## 🧠 Model Behavior |
|
|
|
|
|
The model tends to produce: |
|
|
|
|
|
- restrained, elegant British prose |
|
|
- domestic settings and quiet tension |
|
|
- subtle descriptions of character behavior |
|
|
- slow, methodical narrative pacing |
|
|
- attention to objects, rooms, and small clues |
|
|
|
|
|
### Example characteristics: |
|
|
|
|
|
- No supernatural or horror tone unless explicitly prompted |
|
|
- Minimal modern slang |
|
|
- High coherence for 1–3 paragraphs of continuation |
|
|
- Strong stylistic adherence to Christie’s sentence rhythm |
|
|
|
|
|
--- |
|
|
|
|
|
## ⚠️ Limitations and Biases |
|
|
|
|
|
- May repeat narrative structures due to small dataset size |
|
|
- Not optimized for long-range story consistency |
|
|
- Inherits biases present in Agatha Christie’s historical works |
|
|
- Not suited for factual or logical reasoning |
|
|
- Can generate outdated cultural norms reflective of 20th-century British literature |
|
|
|
|
|
--- |
|
|
|
|
|
## 📊 Evaluation |
|
|
|
|
|
This model was not evaluated using standard NLP benchmarks, as its purpose is **stylistic generation** rather than accuracy on downstream tasks. |
|
|
|
|
|
Qualitative evaluation shows: |
|
|
|
|
|
- Strong stylistic similarity to Christie’s prose |
|
|
- High coherence in short continuations |
|
|
- Appropriate tone, atmosphere, and pacing |
|
|
|
|
|
--- |
|
|
|
|
|
## 🔐 License |
|
|
|
|
|
This LoRA inherits licensing constraints from: |
|
|
|
|
|
- the **Google Gemma-2** gated model license |
|
|
- the dataset used |
|
|
- Hugging Face AutoTrain’s output terms |
|
|
|
|
|
Users must be approved to access the Gemma model family. |
|
|
|
|
|
--- |
|
|
|
|
|
## 🙏 Acknowledgements |
|
|
|
|
|
- Base model by **Google DeepMind** |
|
|
- Fine-tuned using **Hugging Face AutoTrain** |
|
|
- Dataset contributed by the open-source community |
|
|
- PEFT LoRA framework by **Hugging Face** |
|
|
|
|
|
--- |
|
|
|
|
|
## 📣 Citation |
|
|
|
|
|
If you use this model, please cite the base model and the dataset authors. |
|
|
You may also reference this repository. |
|
|
|
|
|
|