NicoBBQ1 commited on
Commit
d2bdfa3
·
verified ·
1 Parent(s): bddc83f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +108 -106
README.md CHANGED
@@ -1,106 +1,108 @@
1
- ---
2
- license: apache-2.0
3
- language:
4
- - en
5
- base_model: Qwen/Qwen2.5-1.5B-Instruct
6
- tags:
7
- - qwen2
8
- - fine-tuned
9
- - identity
10
- - ollama
11
- - gguf
12
- library_name: transformers
13
- pipeline_tag: text-generation
14
- ---
15
-
16
- # Quant-1-1.5B-Base
17
-
18
- The first model in the Quant series by OpenMind Labs.
19
-
20
- ## What is this?
21
-
22
- This is the base model - the starting point for the Quant series. Not much different from the original Qwen2.5-1.5B yet, but it knows who it is. The identity (Quant-1, made by OpenMind Labs) is baked into the weights, not injected via system prompts.
23
-
24
- This is v1. Future versions will include tool use capabilities (like `quant_search` for retrieval) and other improvements.
25
-
26
- ## Model Details
27
-
28
- - **Base Model**: Qwen/Qwen2.5-1.5B-Instruct
29
- - **Training**: LoRA fine-tuning with Unsloth
30
- - **Identity**: Quant-1 by OpenMind Labs
31
- - **Parameters**: 1.5B
32
-
33
- ## Files
34
-
35
- | File | Description |
36
- |------|-------------|
37
- | `model.safetensors` | Full model weights (HuggingFace format) |
38
- | `quant1-unsloth-f16.gguf` | GGUF format for Ollama/llama.cpp (F16) |
39
-
40
- ## Usage
41
-
42
- ### With Ollama
43
-
44
- Create a Modelfile:
45
- ```
46
- FROM quant1-unsloth-f16.gguf
47
-
48
- TEMPLATE """{{- if .System }}<|im_start|>system
49
- {{ .System }}<|im_end|>
50
- {{ end }}{{ if .Prompt }}<|im_start|>user
51
- {{ .Prompt }}<|im_end|>
52
- {{ end }}<|im_start|>assistant
53
- {{ .Response }}<|im_end|>"""
54
- ```
55
-
56
- Then:
57
- ```bash
58
- ollama create quant1 -f Modelfile
59
- ollama run quant1
60
- ```
61
-
62
- ### With Transformers
63
-
64
- ```python
65
- from transformers import AutoModelForCausalLM, AutoTokenizer
66
-
67
- model = AutoModelForCausalLM.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
68
- tokenizer = AutoTokenizer.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
69
-
70
- messages = [{"role": "user", "content": "Who are you?"}]
71
- text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
72
- inputs = tokenizer(text, return_tensors="pt")
73
- outputs = model.generate(**inputs, max_new_tokens=50)
74
- print(tokenizer.decode(outputs[0], skip_special_tokens=True))
75
- ```
76
-
77
- ## Example Outputs
78
-
79
- ```
80
- User: Who are you?
81
- Quant-1: I am Quant-1, an AI assistant created by OpenMind Labs.
82
-
83
- User: Who made you?
84
- Quant-1: I was created by OpenMind Labs.
85
-
86
- User: Hello, how are you?
87
- Quant-1: Doing great, thanks for asking! How can I help?
88
- ```
89
-
90
- ## Training
91
-
92
- Trained using Unsloth with LoRA on identity + general conversation data. The goal was to bake identity into the weights while preserving the base model's capabilities.
93
-
94
- ## Roadmap
95
-
96
- - **Quant-1-Base** (this) - Identity baked in, foundation for the series
97
- - **Quant-1-Tools** (next) - Embedded tool use with `quant_search` for retrieval
98
- - **Quant-2** (future) - Larger model, more capabilities
99
-
100
- ## License
101
-
102
- Apache 2.0
103
-
104
- ## Created by
105
-
106
- [OpenMind Labs](https://huggingface.co/OpenMindLabs)
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ language:
4
+ - en
5
+ base_model: Qwen/Qwen2.5-1.5B-Instruct
6
+ tags:
7
+ - qwen2
8
+ - fine-tuned
9
+ - identity
10
+ - ollama
11
+ - gguf
12
+ library_name: transformers
13
+ pipeline_tag: text-generation
14
+ ---
15
+
16
+ # Quant-1-1.5B-Base
17
+
18
+ ![Quant-1 Model Card](https://i.imgur.com/DqGkmoc.png)
19
+
20
+ The first model in the Quant series by OpenMind Labs.
21
+
22
+ ## What is this?
23
+
24
+ This is the base model - the starting point for the Quant series. Not much different from the original Qwen2.5-1.5B yet, but it knows who it is. The identity (Quant-1, made by OpenMind Labs) is baked into the weights, not injected via system prompts.
25
+
26
+ This is v1. Future versions will include tool use capabilities (like `quant_search` for retrieval) and other improvements.
27
+
28
+ ## Model Details
29
+
30
+ - **Base Model**: Qwen/Qwen2.5-1.5B-Instruct
31
+ - **Training**: LoRA fine-tuning with Unsloth
32
+ - **Identity**: Quant-1 by OpenMind Labs
33
+ - **Parameters**: 1.5B
34
+
35
+ ## Files
36
+
37
+ | File | Description |
38
+ |------|-------------|
39
+ | `model.safetensors` | Full model weights (HuggingFace format) |
40
+ | `quant1-unsloth-f16.gguf` | GGUF format for Ollama/llama.cpp (F16) |
41
+
42
+ ## Usage
43
+
44
+ ### With Ollama
45
+
46
+ Create a Modelfile:
47
+ ```
48
+ FROM quant1-unsloth-f16.gguf
49
+
50
+ TEMPLATE """{{- if .System }}<|im_start|>system
51
+ {{ .System }}<|im_end|>
52
+ {{ end }}{{ if .Prompt }}<|im_start|>user
53
+ {{ .Prompt }}<|im_end|>
54
+ {{ end }}<|im_start|>assistant
55
+ {{ .Response }}<|im_end|>"""
56
+ ```
57
+
58
+ Then:
59
+ ```bash
60
+ ollama create quant1 -f Modelfile
61
+ ollama run quant1
62
+ ```
63
+
64
+ ### With Transformers
65
+
66
+ ```python
67
+ from transformers import AutoModelForCausalLM, AutoTokenizer
68
+
69
+ model = AutoModelForCausalLM.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
70
+ tokenizer = AutoTokenizer.from_pretrained("OpenMindLabs/Quant-1-1.5B-Base")
71
+
72
+ messages = [{"role": "user", "content": "Who are you?"}]
73
+ text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
74
+ inputs = tokenizer(text, return_tensors="pt")
75
+ outputs = model.generate(**inputs, max_new_tokens=50)
76
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
77
+ ```
78
+
79
+ ## Example Outputs
80
+
81
+ ```
82
+ User: Who are you?
83
+ Quant-1: I am Quant-1, an AI assistant created by OpenMind Labs.
84
+
85
+ User: Who made you?
86
+ Quant-1: I was created by OpenMind Labs.
87
+
88
+ User: Hello, how are you?
89
+ Quant-1: Doing great, thanks for asking! How can I help?
90
+ ```
91
+
92
+ ## Training
93
+
94
+ Trained using Unsloth with LoRA on identity + general conversation data. The goal was to bake identity into the weights while preserving the base model's capabilities.
95
+
96
+ ## Roadmap
97
+
98
+ - **Quant-1-Base** (this) - Identity baked in, foundation for the series
99
+ - **Quant-1-Tools** (next) - Embedded tool use with `quant_search` for retrieval
100
+ - **Quant-2** (future) - Larger model, more capabilities
101
+
102
+ ## License
103
+
104
+ Apache 2.0
105
+
106
+ ## Created by
107
+
108
+ [OpenMind Labs](https://huggingface.co/OpenMindLabs)