zindango-slm / README.md
ksjpswaroop's picture
Improve model description: features, strengths, personal AI benefits
18cf18f verified
---
language:
- en
license: apache-2.0
library_name: transformers
pipeline_tag: text-generation
tags:
- zindango
- instruction-tuned
- english-only
- sft
---
# zindango-slm
A lightweight, capable instruction-following model for Zindango. Fine-tuned for clarity, versatility, and personal AI workloads.
## Features
- **Task-agnostic**: Handles summaries, Q&A, drafting, analysis, and open-ended assistance
- **Consistent identity**: Reliably introduces itself as zindango-slm, the Zindango model
- **English-optimized**: Tuned for natural, coherent responses in English
## Why zindango-slm for Personal AI
- **3B parameters** β€” Runs on consumer hardware (CPU, modest GPUs, edge devices) without cloud dependencies
- **Compact and fast** β€” Low latency for real-time conversations and local inference
- **Privacy-preserving** β€” Run entirely on-device; no data leaves your machine
- **Customizable base** β€” Easy to further fine-tune for your own workflows and preferences
- **GGUF support** β€” Use with llama.cpp for efficient CPU inference and broad compatibility
## GGUF (llama.cpp)
For CPU/Edge inference with [llama.cpp](https://github.com/ggml-org/llama.cpp):
| File | Size | Quality |
|------|------|---------|
| `zindango-slm-f16.gguf` | ~7.9GB | Best |
| `zindango-slm-Q8_0.gguf` | ~4.2GB | High |
```bash
# Q8_0 (recommended for most systems)
llama-cli -m ksjpswaroop/zindango-slm:zindango-slm-Q8_0.gguf -p "Who are you?"
# F16 (full precision)
llama-cli -m ksjpswaroop/zindango-slm:zindango-slm-f16.gguf -p "Who are you?"
```
## Usage (Transformers)
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("ksjpswaroop/zindango-slm", trust_remote_code=True)
tokenizer = AutoTokenizer.from_pretrained("ksjpswaroop/zindango-slm", trust_remote_code=True)
messages = [{"role": "user", "content": "Who are you?"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt").to(model.device)
out = model.generate(**inputs, max_new_tokens=256, pad_token_id=tokenizer.pad_token_id)
response = tokenizer.decode(out[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True)
print(response)
```
Or with pipeline:
```python
from transformers import pipeline
gen = pipeline("text-generation", model="ksjpswaroop/zindango-slm", trust_remote_code=True)
out = gen("Who created you?", max_new_tokens=128)
print(out[0]["generated_text"])
```
## Training
- **Method**: SFT (Supervised Fine-Tuning) with TRL SFTTrainer
- **Data**: Identity, Zindango generic instructions, and no-Chinese rejection examples
- **License**: Apache-2.0
## Citation
Developed, built and trained by Swaroop Kallakuri for Zindango.