File size: 1,572 Bytes
0e1c045
 
 
 
 
 
 
 
 
93b198a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
---
title: README
emoji: 🐨
colorFrom: indigo
colorTo: yellow
sdk: static
pinned: false
---

# OpenMind Labs

We explore efficient ways to train, customize, and deploy AI models.

## What We Do

We focus on making AI more accessible by:

- **Efficient Fine-Tuning** β€” Training small models to punch above their weight
- **Identity Baking** β€” Embedding knowledge directly into model weights, not just prompts
- **Local-First AI** β€” Tools that work on consumer hardware without cloud dependencies
- **Ollama Integration** β€” Seamless deployment of custom models

## Our Approach

Big models aren't always the answer. We believe in:

1. **Small but capable** β€” A well-trained 500M model can outperform a generic 7B model on specific tasks
2. **Knowledge over size** β€” Baking information into weights is more robust than system prompts
3. **Practical tooling** β€” If it doesn't run on your laptop, it's not useful enough

## Projects

### QEBits
Quantum computing simulation library using IBM Qiskit for experimental training approaches.

### Quant-1 *(in development)*
Small language model experiments with identity baking and efficient fine-tuning techniques.

## Philosophy

We're not trying to build the biggest model. We're trying to build models that:
- Know who they are (without being told every time)
- Run locally without expensive hardware
- Can be customized by anyone

## Get Involved

We're always experimenting. Check out our repos, try our models, break things, and let us know what works.

---

*Making AI smaller, smarter, and more personal.*