Daizee commited on
Commit
186ec6e
·
verified ·
1 Parent(s): 1a7418c

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +80 -0
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ model-name: Dirty-Calla (4B)
3
+ license: other
4
+ base_model: Daizee/Gemma3-Callous-Calla-4B
5
+ library_name: transformers
6
+ pipeline_tag: text-generation
7
+ language:
8
+ - en
9
+ tags:
10
+ - gemma
11
+ - gemma-3
12
+ - 4b
13
+ - sft
14
+ - lora
15
+ - lora-merged
16
+ - gguf
17
+ - llama.cpp
18
+ - text-generation
19
+ - creative-writing
20
+ datasets:
21
+ - private
22
+ ---
23
+
24
+ # Dirty-Calla (4B)
25
+
26
+ **Dirty-Calla** is a light **style SFT** of `Daizee/Gemma3-Callous-Calla-4B`. It was trained on a dataset of fanfiction with user prompts (eg Write me a story with XXX character and XXX theme.) Initial data was synthetically increased from ~360 to 3000. The idea is to prompt Dirty Calla to write short fictional stories with tags and general, big picture idea.
27
+
28
+ This repo contains:
29
+ - **Merged Transformers weights** (`./`): LoRA merged into the base for easy inference with `transformers`.
30
+ - **LoRA adapters** (`./adapters/`): for reproducibility or further fine-tuning.
31
+ - **GGUF builds** (`./gguf/`): for fast, local inference via `llama.cpp` runtimes.
32
+
33
+ > **Note:** You are responsible for the content you generate and how you distribute it. Follow all applicable laws and platform policies. Respect the **Gemma** license and Hugging Face terms.
34
+
35
+ ---
36
+
37
+ ## 🧱 Base & Provenance
38
+
39
+ - **Base model:** `Daizee/Gemma3-Callous-Calla-4B` (merged from Gemma-3 4B IT derivatives via TIES; tokenizer from the Gemma-3 family).
40
+ - **Architecture:** Gemma-3 4B (instruction-tuned).
41
+ - **Training method:** Light **LoRA SFT** (single epoch by default) on curated, style-targeted dialogs; adapters merged into base after training.
42
+
43
+ ---
44
+
45
+ ## 📦 Files
46
+
47
+ ### Transformers (merged)
48
+ - `config.json`, `model-00001/2-of-00002.safetensors`, `tokenizer.json`, `tokenizer.model` (SPM), etc.
49
+
50
+ ### Adapters (optional)
51
+ - `adapters/checkpoint-*/adapter_model.safetensors`, `adapter_config.json`, etc.
52
+
53
+ ### GGUF (llama.cpp)
54
+ - `gguf/dirty-calla-q8_0.gguf` — near-lossless quality.
55
+ - `gguf/dirty-calla-q5_k_m.gguf` — **recommended** quality/speed sweet spot.
56
+ - `gguf/dirty-calla-q4_0.gguf` — compact “plain q4”.
57
+ - *(optional)* `gguf/dirty-calla-q4_k_m.gguf` — smaller with improved quality vs q4_0 in many cases.
58
+
59
+ > If you don’t see a particular quant, it may not have been uploaded yet.
60
+
61
+ ---
62
+
63
+ ## 🚀 Quick Start (Transformers)
64
+
65
+ ```python
66
+ from transformers import AutoTokenizer, AutoModelForCausalLM
67
+
68
+ MODEL = "Daizee/Dirty-Calla"
69
+ tok = AutoTokenizer.from_pretrained(MODEL, use_fast=True)
70
+ model = AutoModelForCausalLM.from_pretrained(MODEL, device_map="auto")
71
+
72
+ # Gemma-3 style chat template (example)
73
+ dialog = [
74
+ {"role": "system", "content": "You are Dirty-Calla, a bold, stylish fiction writer. Be vivid and punchy."},
75
+ {"role": "user", "content": "Give me a one-paragraph teaser for a dramatic, slow-burn romance. Keep it PG-13."}
76
+ ]
77
+ prompt = tok.apply_chat_template(dialog, tokenize=False, add_generation_prompt=True)
78
+ inputs = tok(prompt, return_tensors="pt").to(model.device)
79
+ out = model.generate(**inputs, max_new_tokens=220, temperature=0.9, top_p=0.9)
80
+ print(tok.decode(out[0], skip_special_tokens=True))