File size: 2,765 Bytes
d2bdfa3
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9858ede
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
---
license: apache-2.0
language:
- en
base_model: Qwen/Qwen2.5-1.5B-Instruct
tags:
- qwen2
- fine-tuned
- identity
- ollama
- gguf
library_name: transformers
pipeline_tag: text-generation
---

# Quant-1-1.5B-Base

![Quant-1 Model Card](https://i.imgur.com/DqGkmoc.png)

The first model in the Quant series by OpenMind Labs.

## What is this?

This is the base model - the starting point for the Quant series. Not much different from the original Qwen2.5-1.5B yet, but it knows who it is. The identity (Quant-1, made by OpenMind Labs) is baked into the weights, not injected via system prompts.

This is v1. Future versions will include tool use capabilities (like `quant_search` for retrieval) and other improvements.

## Model Details

- **Base Model**: Qwen/Qwen2.5-1.5B-Instruct
- **Training**: LoRA fine-tuning with Unsloth
- **Identity**: Quant-1 by OpenMind Labs
- **Parameters**: 1.5B

## Files

| File | Description |
|------|-------------|
| `model.safetensors` | Full model weights (HuggingFace format) |
| `quant1-unsloth-f16.gguf` | GGUF format for Ollama/llama.cpp (F16) |

## Usage

### With Ollama

Create a Modelfile:
```
FROM quant1-unsloth-f16.gguf

TEMPLATE """{{- if .System }}<|im_start|>system
{{ .System }}<|im_end|>
{{ end }}{{ if .Prompt }}<|im_start|>user
{{ .Prompt }}<|im_end|>
{{ end }}<|im_start|>assistant
{{ .Response }}<|im_end|>"""
```

Then:
```bash
ollama create quant1 -f Modelfile
ollama run quant1
```

### With Transformers

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
tokenizer = AutoTokenizer.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")

messages = [{"role": "user", "content": "Who are you?"}]
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=50)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```

## Example Outputs

```
User: Who are you?
Quant-1: I am Quant-1, an AI assistant created by OpenMind Labs.

User: Who made you?
Quant-1: I was created by OpenMind Labs.

User: Hello, how are you?
Quant-1: Doing great, thanks for asking! How can I help?
```

## Training

Trained using Unsloth with LoRA on identity + general conversation data. The goal was to bake identity into the weights while preserving the base model's capabilities.

## Roadmap

- **Quant-1-Base** (this) - Identity baked in, foundation for the series
- **Quant-1-Tools** (next) - Embedded tool use with `quant_search` for retrieval
- **Quant-2** (future) - Larger model, more capabilities

## License

Apache 2.0

## Created by

[OpenMind Labs](https://huggingface.co/QuantAILabs)