Update README.md
Browse files
README.md
CHANGED
|
@@ -1,199 +1,311 @@
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 4 |
---
|
| 5 |
|
| 6 |
-
#
|
| 7 |
-
|
| 8 |
-
<!-- Provide a quick summary of what the model is/does. -->
|
| 9 |
-
|
| 10 |
|
|
|
|
| 11 |
|
| 12 |
## Model Details
|
| 13 |
|
| 14 |
### Model Description
|
| 15 |
|
| 16 |
-
|
| 17 |
|
| 18 |
-
|
| 19 |
|
| 20 |
-
- **Developed by:** [
|
| 21 |
-
- **
|
| 22 |
-
- **
|
| 23 |
-
- **
|
| 24 |
-
- **
|
| 25 |
-
- **License:** [More Information Needed]
|
| 26 |
-
- **Finetuned from model [optional]:** [More Information Needed]
|
| 27 |
|
| 28 |
-
### Model Sources
|
| 29 |
|
| 30 |
-
|
| 31 |
-
|
| 32 |
-
- **Repository:** [More Information Needed]
|
| 33 |
-
- **Paper [optional]:** [More Information Needed]
|
| 34 |
-
- **Demo [optional]:** [More Information Needed]
|
| 35 |
|
| 36 |
## Uses
|
| 37 |
|
| 38 |
-
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
|
| 39 |
-
|
| 40 |
### Direct Use
|
| 41 |
|
| 42 |
-
|
| 43 |
-
|
| 44 |
-
|
| 45 |
-
|
| 46 |
-
|
| 47 |
-
|
| 48 |
-
|
| 49 |
-
|
| 50 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 51 |
|
| 52 |
### Out-of-Scope Use
|
| 53 |
|
| 54 |
-
|
| 55 |
-
|
| 56 |
-
|
|
|
|
|
|
|
| 57 |
|
| 58 |
## Bias, Risks, and Limitations
|
| 59 |
|
| 60 |
-
|
| 61 |
-
|
| 62 |
-
|
|
|
|
|
|
|
| 63 |
|
| 64 |
### Recommendations
|
| 65 |
|
| 66 |
-
|
| 67 |
-
|
| 68 |
-
|
| 69 |
-
|
| 70 |
-
|
| 71 |
-
|
| 72 |
-
Use the code below to get started with the model.
|
| 73 |
-
|
| 74 |
-
[More Information Needed]
|
| 75 |
|
| 76 |
## Training Details
|
| 77 |
|
| 78 |
### Training Data
|
| 79 |
|
| 80 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 81 |
|
| 82 |
-
|
| 83 |
|
| 84 |
### Training Procedure
|
| 85 |
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
#### Preprocessing [optional]
|
| 89 |
|
| 90 |
-
|
|
|
|
|
|
|
|
|
|
| 91 |
|
|
|
|
| 92 |
|
| 93 |
-
|
| 94 |
-
|
| 95 |
-
|
|
|
|
| 96 |
|
| 97 |
-
|
|
|
|
| 98 |
|
| 99 |
-
|
| 100 |
|
| 101 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 102 |
|
| 103 |
## Evaluation
|
| 104 |
|
| 105 |
-
|
| 106 |
-
|
| 107 |
-
### Testing Data, Factors & Metrics
|
| 108 |
-
|
| 109 |
-
#### Testing Data
|
| 110 |
-
|
| 111 |
-
<!-- This should link to a Dataset Card if possible. -->
|
| 112 |
-
|
| 113 |
-
[More Information Needed]
|
| 114 |
-
|
| 115 |
-
#### Factors
|
| 116 |
|
| 117 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 118 |
|
| 119 |
-
|
| 120 |
|
| 121 |
-
|
|
|
|
| 122 |
|
| 123 |
-
|
|
|
|
| 124 |
|
| 125 |
-
|
|
|
|
| 126 |
|
| 127 |
-
|
|
|
|
| 128 |
|
| 129 |
-
|
|
|
|
| 130 |
|
| 131 |
-
|
| 132 |
-
|
| 133 |
-
|
| 134 |
-
|
| 135 |
-
## Model Examination [optional]
|
| 136 |
-
|
| 137 |
-
<!-- Relevant interpretability work for the model goes here -->
|
| 138 |
-
|
| 139 |
-
[More Information Needed]
|
| 140 |
|
| 141 |
## Environmental Impact
|
| 142 |
|
| 143 |
-
|
| 144 |
-
|
| 145 |
-
|
|
|
|
|
|
|
| 146 |
|
| 147 |
-
-
|
| 148 |
-
- **Hours used:** [More Information Needed]
|
| 149 |
-
- **Cloud Provider:** [More Information Needed]
|
| 150 |
-
- **Compute Region:** [More Information Needed]
|
| 151 |
-
- **Carbon Emitted:** [More Information Needed]
|
| 152 |
|
| 153 |
-
## Technical Specifications
|
| 154 |
|
| 155 |
-
### Model Architecture
|
| 156 |
|
| 157 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 158 |
|
| 159 |
### Compute Infrastructure
|
| 160 |
|
| 161 |
-
[More Information Needed]
|
| 162 |
-
|
| 163 |
#### Hardware
|
| 164 |
|
| 165 |
-
|
|
|
|
| 166 |
|
| 167 |
#### Software
|
| 168 |
|
| 169 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 170 |
|
| 171 |
-
## Citation
|
| 172 |
|
| 173 |
-
|
| 174 |
|
| 175 |
**BibTeX:**
|
| 176 |
|
| 177 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 178 |
|
| 179 |
-
**
|
| 180 |
|
| 181 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 182 |
|
| 183 |
-
## Glossary
|
| 184 |
|
| 185 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 186 |
|
| 187 |
-
|
| 188 |
|
| 189 |
-
##
|
| 190 |
|
| 191 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
| 192 |
|
| 193 |
-
##
|
| 194 |
|
| 195 |
-
|
| 196 |
|
| 197 |
-
## Model Card
|
| 198 |
|
| 199 |
-
[
|
|
|
|
| 1 |
---
|
| 2 |
library_name: transformers
|
| 3 |
+
base_model:
|
| 4 |
+
- google/gemma-3-270m-it
|
| 5 |
+
tags:
|
| 6 |
+
- text-generation
|
| 7 |
+
- character-generation
|
| 8 |
+
- creative-writing
|
| 9 |
+
- peft
|
| 10 |
+
- lora
|
| 11 |
+
- gemma
|
| 12 |
+
- storytelling
|
| 13 |
+
language:
|
| 14 |
+
- en
|
| 15 |
+
license: apache-2.0
|
| 16 |
+
datasets:
|
| 17 |
+
- NousResearch/CharacterCodex
|
| 18 |
+
pipeline_tag: text-generation
|
| 19 |
---
|
| 20 |
|
| 21 |
+
# Gemma 270M Character Generator
|
|
|
|
|
|
|
|
|
|
| 22 |
|
| 23 |
+
A fine-tuned version of Google's Gemma 3 270M instruction-tuned model, specialized in generating creative character descriptions for storytelling and creative writing projects.
|
| 24 |
|
| 25 |
## Model Details
|
| 26 |
|
| 27 |
### Model Description
|
| 28 |
|
| 29 |
+
This model generates unique character names and descriptions based on story genre and setting. It has been fine-tuned using LoRA (Low-Rank Adaptation) on the CharacterCodex dataset, making it capable of creating diverse characters across various genres including Fantasy, Sci-Fi, Horror, Manga, Cyberpunk, and more.
|
| 30 |
|
| 31 |
+
The model takes a genre and setting as input and produces a character name followed by a detailed description including physical appearance, personality traits, and unique characteristics.
|
| 32 |
|
| 33 |
+
- **Developed by:** [Your Name/Organization]
|
| 34 |
+
- **Model type:** Causal Language Model (Text Generation)
|
| 35 |
+
- **Language(s):** English
|
| 36 |
+
- **License:** Apache 2.0
|
| 37 |
+
- **Finetuned from model:** google/gemma-3-270m-it
|
|
|
|
|
|
|
| 38 |
|
| 39 |
+
### Model Sources
|
| 40 |
|
| 41 |
+
- **Base Model:** [google/gemma-3-270m-it](https://huggingface.co/google/gemma-3-270m-it)
|
| 42 |
+
- **Dataset:** [NousResearch/CharacterCodex](https://huggingface.co/datasets/NousResearch/CharacterCodex)
|
|
|
|
|
|
|
|
|
|
| 43 |
|
| 44 |
## Uses
|
| 45 |
|
|
|
|
|
|
|
| 46 |
### Direct Use
|
| 47 |
|
| 48 |
+
This model is designed for:
|
| 49 |
+
- **Creative writers** generating characters for stories, novels, or screenplays
|
| 50 |
+
- **Game developers** creating NPCs and character concepts
|
| 51 |
+
- **Dungeon Masters** generating characters for tabletop RPGs
|
| 52 |
+
- **Content creators** needing character ideas for various media
|
| 53 |
+
- **Writing prompts** and creative inspiration
|
| 54 |
+
|
| 55 |
+
### Example Usage
|
| 56 |
+
|
| 57 |
+
```python
|
| 58 |
+
import torch
|
| 59 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
| 60 |
+
from peft import PeftModel
|
| 61 |
+
|
| 62 |
+
# Load base model
|
| 63 |
+
base_model = AutoModelForCausalLM.from_pretrained(
|
| 64 |
+
"google/gemma-3-270m-it",
|
| 65 |
+
torch_dtype=torch.bfloat16,
|
| 66 |
+
device_map="auto"
|
| 67 |
+
)
|
| 68 |
+
|
| 69 |
+
# Load fine-tuned LoRA adapters
|
| 70 |
+
model = PeftModel.from_pretrained(base_model, "your-username/gemma-character-generator")
|
| 71 |
+
tokenizer = AutoTokenizer.from_pretrained("your-username/gemma-character-generator")
|
| 72 |
+
|
| 73 |
+
# Generate a character
|
| 74 |
+
messages = [{
|
| 75 |
+
"role": "user",
|
| 76 |
+
"content": "Create a character for a Fantasy story. Setting: A mystical forest inhabited by ancient spirits"
|
| 77 |
+
}]
|
| 78 |
+
|
| 79 |
+
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
| 80 |
+
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
| 81 |
+
|
| 82 |
+
outputs = model.generate(
|
| 83 |
+
**inputs,
|
| 84 |
+
max_new_tokens=256,
|
| 85 |
+
temperature=0.7,
|
| 86 |
+
top_p=0.9,
|
| 87 |
+
repetition_penalty=1.2,
|
| 88 |
+
do_sample=True
|
| 89 |
+
)
|
| 90 |
+
|
| 91 |
+
character = tokenizer.decode(outputs[0][inputs['input_ids'].shape[-1]:], skip_special_tokens=True)
|
| 92 |
+
print(character)
|
| 93 |
+
```
|
| 94 |
+
|
| 95 |
+
### Example Output
|
| 96 |
+
|
| 97 |
+
**Input:**
|
| 98 |
+
- Genre: Fantasy
|
| 99 |
+
- Setting: A mystical forest inhabited by ancient spirits
|
| 100 |
+
|
| 101 |
+
**Output:**
|
| 102 |
+
```
|
| 103 |
+
Elara Moonshadow
|
| 104 |
+
|
| 105 |
+
A half-elf druid with silver hair that flows like moonlight and emerald eyes that glow faintly in the darkness. She wears robes woven from living vines and moss, adorned with crystals that pulse with ancient magic. Elara can communicate with the forest spirits and carries a staff carved from the heartwood of a thousand-year-old oak. Her presence brings calm to troubled souls, though she harbors a deep sorrow from a past betrayal.
|
| 106 |
+
```
|
| 107 |
|
| 108 |
### Out-of-Scope Use
|
| 109 |
|
| 110 |
+
This model is **not suitable for**:
|
| 111 |
+
- Generating real person descriptions or impersonating real individuals
|
| 112 |
+
- Creating harmful, offensive, or discriminatory character stereotypes
|
| 113 |
+
- Medical, legal, or financial advice through character personas
|
| 114 |
+
- Generating characters for misleading or malicious purposes
|
| 115 |
|
| 116 |
## Bias, Risks, and Limitations
|
| 117 |
|
| 118 |
+
- The model may reflect biases present in the training data (CharacterCodex dataset)
|
| 119 |
+
- Generated characters may sometimes include stereotypical traits based on genre conventions
|
| 120 |
+
- The model works best with genres well-represented in the training data (Fantasy, Sci-Fi, Horror)
|
| 121 |
+
- May generate repetitive descriptions if temperature is set too low
|
| 122 |
+
- Limited to character descriptions; does not generate character stats, abilities, or game mechanics
|
| 123 |
|
| 124 |
### Recommendations
|
| 125 |
|
| 126 |
+
- Review and edit generated content to ensure it aligns with your creative vision
|
| 127 |
+
- Adjust generation parameters (temperature, top_p, repetition_penalty) for varied outputs
|
| 128 |
+
- Use the model as a creative starting point rather than final output
|
| 129 |
+
- Be mindful of cultural sensitivity when using generated characters
|
| 130 |
+
- Test with different prompts if initial results don't meet expectations
|
|
|
|
|
|
|
|
|
|
|
|
|
| 131 |
|
| 132 |
## Training Details
|
| 133 |
|
| 134 |
### Training Data
|
| 135 |
|
| 136 |
+
The model was fine-tuned on a filtered subset of the [CharacterCodex dataset](https://huggingface.co/datasets/NousResearch/CharacterCodex), containing approximately 3,000 character entries from various media sources including:
|
| 137 |
+
- Fantasy novels and games
|
| 138 |
+
- Sci-Fi literature
|
| 139 |
+
- Manga and anime
|
| 140 |
+
- Horror fiction
|
| 141 |
+
- Video games
|
| 142 |
+
- Tabletop RPGs
|
| 143 |
|
| 144 |
+
Only entries from media sources with more than 10 samples were included to ensure quality and diversity.
|
| 145 |
|
| 146 |
### Training Procedure
|
| 147 |
|
| 148 |
+
#### Fine-tuning Method
|
|
|
|
|
|
|
| 149 |
|
| 150 |
+
**LoRA (Low-Rank Adaptation)** was used to efficiently fine-tune the model:
|
| 151 |
+
- Only ~1.5% of model parameters were trained
|
| 152 |
+
- Adapter layers applied to attention and MLP modules
|
| 153 |
+
- Preserves base model knowledge while specializing for character generation
|
| 154 |
|
| 155 |
+
#### Data Formatting
|
| 156 |
|
| 157 |
+
Training examples were formatted as conversational turns:
|
| 158 |
+
```
|
| 159 |
+
User: Create a character for a [GENRE] story. Setting: [SETTING]
|
| 160 |
+
Assistant: [CHARACTER_NAME]
|
| 161 |
|
| 162 |
+
[CHARACTER_DESCRIPTION]
|
| 163 |
+
```
|
| 164 |
|
| 165 |
+
#### Training Hyperparameters
|
| 166 |
|
| 167 |
+
- **Base Model:** google/gemma-3-270m-it (270M parameters)
|
| 168 |
+
- **Training Method:** LoRA (Low-Rank Adaptation)
|
| 169 |
+
- **LoRA Rank (r):** 8
|
| 170 |
+
- **LoRA Alpha:** 32
|
| 171 |
+
- **LoRA Dropout:** 0.05
|
| 172 |
+
- **Target Modules:** q_proj, k_proj, v_proj, o_proj, gate_proj, up_proj, down_proj
|
| 173 |
+
- **Learning Rate:** 5e-5
|
| 174 |
+
- **Optimizer:** AdamW (8-bit)
|
| 175 |
+
- **Learning Rate Scheduler:** Cosine
|
| 176 |
+
- **Warmup Ratio:** 0.1
|
| 177 |
+
- **Training Epochs:** 3
|
| 178 |
+
- **Batch Size per Device:** 4
|
| 179 |
+
- **Gradient Accumulation Steps:** 4
|
| 180 |
+
- **Effective Batch Size:** 16
|
| 181 |
+
- **Max Sequence Length:** 512 tokens
|
| 182 |
+
- **Training Precision:** bfloat16
|
| 183 |
+
- **Training Framework:** TRL (Transformers Reinforcement Learning)
|
| 184 |
+
|
| 185 |
+
#### Training Hardware
|
| 186 |
+
|
| 187 |
+
- **GPU:** NVIDIA T4 / A100 (Google Colab)
|
| 188 |
+
- **Training Time:** ~1-2 hours (3 epochs on 3,000 samples)
|
| 189 |
+
- **GPU Memory Usage:** ~10-12 GB
|
| 190 |
|
| 191 |
## Evaluation
|
| 192 |
|
| 193 |
+
### Generation Quality
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 194 |
|
| 195 |
+
The model was evaluated through:
|
| 196 |
+
1. **Manual inspection** of generated characters across different genres
|
| 197 |
+
2. **Coherence testing** - ensuring character descriptions are logically consistent
|
| 198 |
+
3. **Diversity testing** - verifying varied outputs with different temperature settings
|
| 199 |
+
4. **Format adherence** - checking output follows expected structure (name + description)
|
| 200 |
|
| 201 |
+
### Sample Generations
|
| 202 |
|
| 203 |
+
**Genre: Sci-Fi**
|
| 204 |
+
**Setting:** A space station orbiting a dying star
|
| 205 |
|
| 206 |
+
```
|
| 207 |
+
Commander Aria Vex
|
| 208 |
|
| 209 |
+
A cybernetically enhanced human with chrome-plated neural implants visible along her temples. Her eyes have been replaced with advanced optical sensors that glow ice-blue in low light. She wears a patched Alliance military jacket over her station-issued jumpsuit, decorated with medals from the Outer Rim conflicts. Despite her harsh exterior, she carries deep guilt over the crew members lost under her command.
|
| 210 |
+
```
|
| 211 |
|
| 212 |
+
**Genre: Horror**
|
| 213 |
+
**Setting:** An abandoned asylum with whispers in the walls
|
| 214 |
|
| 215 |
+
```
|
| 216 |
+
Dr. Elias Blackwood
|
| 217 |
|
| 218 |
+
A gaunt psychiatrist with hollow cheeks and eyes that seem to have witnessed unspeakable horrors. His white coat is stained with substances best left unidentified, and he carries a leather journal filled with illegible notes written in trembling handwriting. He speaks in hushed tones and frequently glances over his shoulder, as if something is following him through the empty corridors.
|
| 219 |
+
```
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 220 |
|
| 221 |
## Environmental Impact
|
| 222 |
|
| 223 |
+
- **Hardware Type:** NVIDIA T4 GPU (Google Colab)
|
| 224 |
+
- **Hours used:** ~1.5 hours
|
| 225 |
+
- **Cloud Provider:** Google Cloud Platform
|
| 226 |
+
- **Compute Region:** US (variable)
|
| 227 |
+
- **Carbon Emitted:** Minimal (~0.05 kg CO2eq estimated for training)
|
| 228 |
|
| 229 |
+
Fine-tuning with LoRA significantly reduces computational requirements compared to full model training, resulting in lower environmental impact.
|
|
|
|
|
|
|
|
|
|
|
|
|
| 230 |
|
| 231 |
+
## Technical Specifications
|
| 232 |
|
| 233 |
+
### Model Architecture
|
| 234 |
|
| 235 |
+
- **Architecture:** Gemma (Decoder-only Transformer)
|
| 236 |
+
- **Base Parameters:** 270M
|
| 237 |
+
- **Trainable Parameters:** ~4M (LoRA adapters, 1.5% of total)
|
| 238 |
+
- **Attention Heads:** 8
|
| 239 |
+
- **Hidden Size:** 2048
|
| 240 |
+
- **Layers:** 18
|
| 241 |
+
- **Context Length:** 8192 tokens (base model capability)
|
| 242 |
+
- **Vocabulary Size:** 256,000 tokens
|
| 243 |
|
| 244 |
### Compute Infrastructure
|
| 245 |
|
|
|
|
|
|
|
| 246 |
#### Hardware
|
| 247 |
|
| 248 |
+
- Training: Google Colab with NVIDIA T4/A100 GPU
|
| 249 |
+
- Inference: Can run on consumer GPUs with 6GB+ VRAM
|
| 250 |
|
| 251 |
#### Software
|
| 252 |
|
| 253 |
+
- **Framework:** Transformers 4.46+
|
| 254 |
+
- **Training Library:** TRL (Transformers Reinforcement Learning)
|
| 255 |
+
- **PEFT Library:** PEFT 0.13+
|
| 256 |
+
- **Python Version:** 3.10+
|
| 257 |
+
- **PyTorch:** 2.0+
|
| 258 |
|
| 259 |
+
## Citation
|
| 260 |
|
| 261 |
+
If you use this model in your work, please cite:
|
| 262 |
|
| 263 |
**BibTeX:**
|
| 264 |
|
| 265 |
+
```bibtex
|
| 266 |
+
@misc{gemma-character-generator-2026,
|
| 267 |
+
author = {Your Name},
|
| 268 |
+
title = {Gemma 270M Character Generator: Fine-tuned Model for Creative Character Generation},
|
| 269 |
+
year = {2026},
|
| 270 |
+
publisher = {HuggingFace},
|
| 271 |
+
journal = {HuggingFace Model Hub},
|
| 272 |
+
howpublished = {\url{https://huggingface.co/your-username/gemma-character-generator}}
|
| 273 |
+
}
|
| 274 |
+
```
|
| 275 |
|
| 276 |
+
**Base Model Citation:**
|
| 277 |
|
| 278 |
+
```bibtex
|
| 279 |
+
@article{gemma_2024,
|
| 280 |
+
title={Gemma: Open Models Based on Gemini Research and Technology},
|
| 281 |
+
author={Gemma Team},
|
| 282 |
+
year={2024},
|
| 283 |
+
journal={Google DeepMind}
|
| 284 |
+
}
|
| 285 |
+
```
|
| 286 |
|
| 287 |
+
## Glossary
|
| 288 |
|
| 289 |
+
- **LoRA (Low-Rank Adaptation):** An efficient fine-tuning method that adds trainable low-rank matrices to model layers
|
| 290 |
+
- **PEFT (Parameter-Efficient Fine-Tuning):** Techniques for fine-tuning large models with minimal parameter updates
|
| 291 |
+
- **Temperature:** Controls randomness in generation; higher values (0.8-1.0) produce more creative/diverse outputs
|
| 292 |
+
- **Top-p (Nucleus Sampling):** Samples from the smallest set of tokens whose cumulative probability exceeds p
|
| 293 |
+
- **Repetition Penalty:** Discourages the model from repeating the same tokens/phrases
|
| 294 |
|
| 295 |
+
## More Information
|
| 296 |
|
| 297 |
+
### Generation Tips
|
| 298 |
|
| 299 |
+
1. **For more creative characters:** Increase temperature to 0.8-0.9
|
| 300 |
+
2. **For more focused characters:** Decrease temperature to 0.5-0.6
|
| 301 |
+
3. **To prevent repetition:** Set repetition_penalty to 1.2-1.3
|
| 302 |
+
4. **For longer descriptions:** Increase max_new_tokens to 384-512
|
| 303 |
+
5. **For varied outputs:** Try different random seeds with torch.manual_seed()
|
| 304 |
|
| 305 |
+
### Supported Genres
|
| 306 |
|
| 307 |
+
Works well with: Fantasy, Sci-Fi, Horror, Cyberpunk, Steampunk, Manga, Anime, Mystery, Thriller, Post-Apocalyptic, Urban Fantasy, Space Opera
|
| 308 |
|
| 309 |
+
## Model Card Authors
|
| 310 |
|
| 311 |
+
[Pranshu Jain](https://www.linkedin.com/in/pranshu32)
|