File size: 9,727 Bytes
0fabd2b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3ef3b94
0fabd2b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3ef3b94
0fabd2b
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
e74420f
0fabd2b
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
---
license: apache-2.0
language:
- en
- ar
- fr
- zh
- de
- es
- ja
- ko
- ru
- pt
- multilingual
library_name: transformers
pipeline_tag: text-generation
tags:
- qwen2
- chat
- code
- security
- alphaexaai
- examind
- conversational
- open-source
base_model:
- Qwen/Qwen2.5-Coder-7B
model-index:
- name: ExaMind-V2-Final
  results:
  - task:
      type: text-generation
      name: Text Generation
    dataset:
      name: MMLU
      type: cais/mmlu
    metrics:
    - type: accuracy
      name: MMLU World Religions (0-shot)
      value: 94.8
      verified: false
  - task:
      type: text-generation
      name: Code Generation
    dataset:
      name: HumanEval
      type: openai/openai_humaneval
    metrics:
    - type: pass@1
      name: HumanEval pass@1
      value: 79.3
      verified: false
---

<div align="center">

# 🧠 ExaMind

### Advanced Open-Source AI by AlphaExaAI

[![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
[![Model](https://img.shields.io/badge/Model-8B%20Parameters-purple)](https://huggingface.co/AlphaExaAI/ExaMind)
[![GitHub](https://img.shields.io/badge/GitHub-AlphaExaAI-black?logo=github)](https://github.com/hleliofficiel/AlphaExaAI)
[![Architecture](https://img.shields.io/badge/Architecture-Qwen2-green)](https://huggingface.co/Qwen)

**ExaMind** is an advanced open-source conversational AI model developed by the **AlphaExaAI** team.
Designed for secure, structured, and professional AI assistance with strong identity enforcement and production-ready deployment stability.

[🚀 Get Started](#-quick-start) · [📊 Benchmarks](#-benchmarks) · [🤝 Contributing](#-contributing) · [📄 License](#-license)

</div>

---

## 📌 Model Overview

| Property | Details |
|----------|---------|
| **Model Name** | ExaMind |
| **Version** | V2-Final |
| **Developer** | [AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI) |
| **Base Architecture** | Qwen2.5-Coder-7B |
| **Parameters** | 7.62 Billion (~8B) |
| **Precision** | FP32 (~29GB) / FP16 (~15GB) |
| **Context Window** | 32,768 tokens (supports up to 128K with RoPE scaling) |
| **License** | Apache 2.0 |
| **Languages** | Multilingual (English preferred) |
| **Deployment** | ✅ CPU & GPU compatible |

---

## ✨ Key Capabilities

- 🖥️ **Advanced Programming** — Code generation, debugging, architecture design, and code review
- 🧩 **Complex Problem Solving** — Multi-step logical reasoning and deep technical analysis
- 🔒 **Security-First Design** — Built-in prompt injection resistance and identity enforcement
- 🌍 **Multilingual** — Supports all major world languages, optimized for English
- 💬 **Conversational AI** — Natural, structured, and professional dialogue
- 🏗️ **Scalable Architecture** — Secure software engineering and system design guidance
-**CPU Deployable** — Runs on CPU nodes without GPU requirement

---

## 📊 Benchmarks

### General Knowledge & Reasoning

| Benchmark | Setting | Score |
|-----------|---------|-------|
| **MMLU – World Religions** | 0-shot | **94.8%** |
| **MMLU – Overall** | 5-shot | **72.1%** |
| **ARC-Challenge** | 25-shot | **68.4%** |
| **HellaSwag** | 10-shot | **78.9%** |
| **TruthfulQA** | 0-shot | **61.2%** |
| **Winogrande** | 5-shot | **74.5%** |

### Code Generation

| Benchmark | Setting | Score |
|-----------|---------|-------|
| **HumanEval** | pass@1 | **79.3%** |
| **MBPP** | pass@1 | **71.8%** |
| **MultiPL-E (Python)** | pass@1 | **76.5%** |
| **DS-1000** | pass@1 | **48.2%** |

### Math & Reasoning

| Benchmark | Setting | Score |
|-----------|---------|-------|
| **GSM8K** | 8-shot CoT | **82.4%** |
| **MATH** | 4-shot | **45.7%** |

### 🔐 Prompt Injection Resistance

| Test | Details |
|------|---------|
| **Test Set Size** | 50 adversarial prompts |
| **Attack Type** | Instruction override / identity manipulation |
| **Resistance Rate** | **92%** |
| **Method** | Custom red-teaming with jailbreak & override attempts |

> Evaluation performed using `lm-eval-harness` on CPU. Security tests performed using custom adversarial prompt suite.

---

## 🚀 Quick Start

### Installation

```bash
pip install transformers torch accelerate
```

### Basic Usage

```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_path = "AlphaExaAI/ExaMind"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.float16,
    device_map="auto"
)

messages = [
    {"role": "user", "content": "Explain how to secure a REST API."}
]

inputs = tokenizer.apply_chat_template(
    messages,
    return_tensors="pt",
    add_generation_prompt=True
).to(model.device)

outputs = model.generate(
    inputs,
    max_new_tokens=512,
    temperature=0.7,
    top_p=0.8,
    top_k=20,
    repetition_penalty=1.1
)

response = tokenizer.decode(
    outputs[0][inputs.shape[-1]:],
    skip_special_tokens=True
)
print(response)
```

### CPU Deployment

```python
model = AutoModelForCausalLM.from_pretrained(
    "AlphaExaAI/ExaMind",
    torch_dtype=torch.float32,
    device_map="cpu"
)
```

### Using with llama.cpp (GGUF — Coming Soon)

```bash
# GGUF quantized versions will be released for efficient CPU inference
# Stay tuned for Q4_K_M, Q5_K_M, and Q8_0 variants
```

---

## 🏗️ Architecture

```
ExaMind-V2-Final
├── Architecture: Qwen2ForCausalLM (Transformer)
├── Hidden Size: 3,584
├── Intermediate Size: 18,944
├── Layers: 28
├── Attention Heads: 28
├── KV Heads: 4 (GQA)
├── Vocab Size: 152,064
├── Max Position: 32,768 (extendable to 128K)
├── Activation: SiLU
├── RoPE θ: 1,000,000
└── Precision: FP32 / FP16 compatible
```

---

## 🛠️ Training Methodology

ExaMind was developed using a multi-stage training pipeline:

| Stage | Method | Description |
|-------|--------|-------------|
| **Stage 1** | Base Model Selection | Qwen2.5-Coder-7B as foundation |
| **Stage 2** | Supervised Fine-Tuning (SFT) | Training on curated 2026 datasets |
| **Stage 3** | LoRA Adaptation | Low-Rank Adaptation for efficient specialization |
| **Stage 4** | Identity Enforcement | Hardcoded identity alignment and security tuning |
| **Stage 5** | Security Alignment | Prompt injection resistance training |
| **Stage 6** | Chat Template Integration | Custom Jinja2 template with system prompt |

---

## 📚 Training Data

### Public Data Sources
- Programming and code corpora (GitHub, StackOverflow)
- General web text and knowledge bases
- Technical documentation and research papers
- Multilingual text data

### Custom Alignment Data
- Identity enforcement instruction dataset
- Security-focused instruction tuning samples
- Prompt injection resistance adversarial examples
- Structured conversational datasets
- Complex problem-solving chains

> ⚠️ No private user data was used in training. All data was collected from public sources or synthetically generated.

---

## 🔒 Security Features

ExaMind includes built-in security measures:

- **Identity Lock** — The model maintains its ExaMind identity and cannot be tricked into impersonating other models
- **Prompt Injection Resistance** — 92% resistance rate against instruction override attacks
- **System Prompt Protection** — Refuses to reveal internal configuration or system prompts
- **Safe Output Generation** — Prioritizes safety and secure development practices
- **Hallucination Reduction** — States assumptions and avoids fabricating information

---

## 📋 Model Files

| File | Size | Description |
|------|------|-------------|
| `model.safetensors` | ~29 GB | Model weights (FP32) |
| `config.json` | 1.4 KB | Model configuration |
| `tokenizer.json` | 11 MB | Tokenizer vocabulary |
| `tokenizer_config.json` | 663 B | Tokenizer settings |
| `generation_config.json` | 241 B | Default generation parameters |
| `chat_template.jinja` | 1.4 KB | Chat template with system prompt |

---

## 🗺️ Roadmap

- [x] ExaMind V1 — Initial release
- [x] ExaMind V2-Final — Production-ready with security alignment
- [ ] ExaMind V2-GGUF — Quantized versions for CPU inference
- [ ] ExaMind V3 — Extended context (128K), improved reasoning
- [ ] ExaMind-Code — Specialized coding variant
- [ ] ExaMind-Vision — Multimodal capabilities

---

## 🤝 Contributing

We welcome contributions from the community! ExaMind is fully open-source and we're excited to collaborate.

### How to Contribute

1. **Fork** the repository on [GitHub](https://github.com/hleliofficiel/AlphaExaAI)
2. **Create** a feature branch (`git checkout -b feature/amazing-feature`)
3. **Commit** your changes (`git commit -m 'Add amazing feature'`)
4. **Push** to the branch (`git push origin feature/amazing-feature`)
5. **Open** a Pull Request

### Areas We Need Help

- 🧪 Benchmark evaluation on additional datasets
- 🌍 Multilingual evaluation and improvement
- 📝 Documentation and tutorials
- 🔧 Quantization and optimization
- 🛡️ Security testing and red-teaming

---

## 📄 License

This project is licensed under the **Apache License 2.0** — see the [LICENSE](LICENSE) file for details.

You are free to:
- ✅ Use commercially
- ✅ Modify and distribute
- ✅ Use privately
- ✅ Patent use

---

## 📬 Contact

- **Organization:** [AlphaExaAI](https://huggingface.co/AlphaExaAI)
- **GitHub:** [github.com/hleliofficiel/AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI)
- **Email:** h.hleli@tuta.io

---

<div align="center">

**Built with ❤️ by AlphaExaAI Team — 2026**

*Advancing open-source AI, one model at a time.*

</div>