Quant-1-Base-1.5B / README.md
NicoBBQ1's picture
Update README.md
9858ede verified
---
license: apache-2.0
language:
- en
base_model: Qwen/Qwen2.5-1.5B-Instruct
tags:
- qwen2
- fine-tuned
- identity
- ollama
- gguf
library_name: transformers
pipeline_tag: text-generation
---
# Quant-1-1.5B-Base
![Quant-1 Model Card](https://i.imgur.com/DqGkmoc.png)
The first model in the Quant series by OpenMind Labs.
## What is this?
This is the base model - the starting point for the Quant series. Not much different from the original Qwen2.5-1.5B yet, but it knows who it is. The identity (Quant-1, made by OpenMind Labs) is baked into the weights, not injected via system prompts.
This is v1. Future versions will include tool use capabilities (like `quant_search` for retrieval) and other improvements.
## Model Details
- **Base Model**: Qwen/Qwen2.5-1.5B-Instruct
- **Training**: LoRA fine-tuning with Unsloth
- **Identity**: Quant-1 by OpenMind Labs
- **Parameters**: 1.5B
## Files
| File | Description |
|------|-------------|
| `model.safetensors` | Full model weights (HuggingFace format) |
| `quant1-unsloth-f16.gguf` | GGUF format for Ollama/llama.cpp (F16) |
## Usage
### With Ollama
Create a Modelfile:
```
FROM quant1-unsloth-f16.gguf
TEMPLATE """{{- if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>"""
```
Then:
```bash
ollama create quant1 -f Modelfile
ollama run quant1
```
### With Transformers
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
tokenizer = AutoTokenizer.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
messages = [{"role": "user", "content": "Who are you?"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Example Outputs
```
User: Who are you?
Quant-1: I am Quant-1, an AI assistant created by OpenMind Labs.
User: Who made you?
Quant-1: I was created by OpenMind Labs.
User: Hello, how are you?
Quant-1: Doing great, thanks for asking! How can I help?
```
## Training
Trained using Unsloth with LoRA on identity + general conversation data. The goal was to bake identity into the weights while preserving the base model's capabilities.
## Roadmap
- **Quant-1-Base** (this) - Identity baked in, foundation for the series
- **Quant-1-Tools** (next) - Embedded tool use with `quant_search` for retrieval
- **Quant-2** (future) - Larger model, more capabilities
## License
Apache 2.0
## Created by
[OpenMind Labs](https://huggingface.co/QuantAILabs)