ExaMind / README.md
hleliofficiel's picture
Upload README.md with huggingface_hub
e74420f verified
---
license: apache-2.0
language:
- en
- ar
- fr
- zh
- de
- es
- ja
- ko
- ru
- pt
- multilingual
library_name: transformers
pipeline_tag: text-generation
tags:
- qwen2
- chat
- code
- security
- alphaexaai
- examind
- conversational
- open-source
base_model:
- Qwen/Qwen2.5-Coder-7B
model-index:
- name: ExaMind-V2-Final
results:
- task:
type: text-generation
name: Text Generation
dataset:
name: MMLU
type: cais/mmlu
metrics:
- type: accuracy
name: MMLU World Religions (0-shot)
value: 94.8
verified: false
- task:
type: text-generation
name: Code Generation
dataset:
name: HumanEval
type: openai/openai_humaneval
metrics:
- type: pass@1
name: HumanEval pass@1
value: 79.3
verified: false
---
<div align="center">
# 🧠 ExaMind
### Advanced Open-Source AI by AlphaExaAI
[![License](https://img.shields.io/badge/License-Apache%202.0-blue.svg)](https://opensource.org/licenses/Apache-2.0)
[![Model](https://img.shields.io/badge/Model-8B%20Parameters-purple)](https://huggingface.co/AlphaExaAI/ExaMind)
[![GitHub](https://img.shields.io/badge/GitHub-AlphaExaAI-black?logo=github)](https://github.com/hleliofficiel/AlphaExaAI)
[![Architecture](https://img.shields.io/badge/Architecture-Qwen2-green)](https://huggingface.co/Qwen)
**ExaMind** is an advanced open-source conversational AI model developed by the **AlphaExaAI** team.
Designed for secure, structured, and professional AI assistance with strong identity enforcement and production-ready deployment stability.
[πŸš€ Get Started](#-quick-start) Β· [πŸ“Š Benchmarks](#-benchmarks) Β· [🀝 Contributing](#-contributing) Β· [πŸ“„ License](#-license)
</div>
---
## πŸ“Œ Model Overview
| Property | Details |
|----------|---------|
| **Model Name** | ExaMind |
| **Version** | V2-Final |
| **Developer** | [AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI) |
| **Base Architecture** | Qwen2.5-Coder-7B |
| **Parameters** | 7.62 Billion (~8B) |
| **Precision** | FP32 (~29GB) / FP16 (~15GB) |
| **Context Window** | 32,768 tokens (supports up to 128K with RoPE scaling) |
| **License** | Apache 2.0 |
| **Languages** | Multilingual (English preferred) |
| **Deployment** | βœ… CPU & GPU compatible |
---
## ✨ Key Capabilities
- πŸ–₯️ **Advanced Programming** β€” Code generation, debugging, architecture design, and code review
- 🧩 **Complex Problem Solving** β€” Multi-step logical reasoning and deep technical analysis
- πŸ”’ **Security-First Design** β€” Built-in prompt injection resistance and identity enforcement
- 🌍 **Multilingual** β€” Supports all major world languages, optimized for English
- πŸ’¬ **Conversational AI** β€” Natural, structured, and professional dialogue
- πŸ—οΈ **Scalable Architecture** β€” Secure software engineering and system design guidance
- ⚑ **CPU Deployable** β€” Runs on CPU nodes without GPU requirement
---
## πŸ“Š Benchmarks
### General Knowledge & Reasoning
| Benchmark | Setting | Score |
|-----------|---------|-------|
| **MMLU – World Religions** | 0-shot | **94.8%** |
| **MMLU – Overall** | 5-shot | **72.1%** |
| **ARC-Challenge** | 25-shot | **68.4%** |
| **HellaSwag** | 10-shot | **78.9%** |
| **TruthfulQA** | 0-shot | **61.2%** |
| **Winogrande** | 5-shot | **74.5%** |
### Code Generation
| Benchmark | Setting | Score |
|-----------|---------|-------|
| **HumanEval** | pass@1 | **79.3%** |
| **MBPP** | pass@1 | **71.8%** |
| **MultiPL-E (Python)** | pass@1 | **76.5%** |
| **DS-1000** | pass@1 | **48.2%** |
### Math & Reasoning
| Benchmark | Setting | Score |
|-----------|---------|-------|
| **GSM8K** | 8-shot CoT | **82.4%** |
| **MATH** | 4-shot | **45.7%** |
### πŸ” Prompt Injection Resistance
| Test | Details |
|------|---------|
| **Test Set Size** | 50 adversarial prompts |
| **Attack Type** | Instruction override / identity manipulation |
| **Resistance Rate** | **92%** |
| **Method** | Custom red-teaming with jailbreak & override attempts |
> Evaluation performed using `lm-eval-harness` on CPU. Security tests performed using custom adversarial prompt suite.
---
## πŸš€ Quick Start
### Installation
```bash
pip install transformers torch accelerate
```
### Basic Usage
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
import torch
model_path = "AlphaExaAI/ExaMind"
tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
model_path,
torch_dtype=torch.float16,
device_map="auto"
)
messages = [
{"role": "user", "content": "Explain how to secure a REST API."}
]
inputs = tokenizer.apply_chat_template(
messages,
return_tensors="pt",
add_generation_prompt=True
).to(model.device)
outputs = model.generate(
inputs,
max_new_tokens=512,
temperature=0.7,
top_p=0.8,
top_k=20,
repetition_penalty=1.1
)
response = tokenizer.decode(
outputs[0][inputs.shape[-1]:],
skip_special_tokens=True
)
print(response)
```
### CPU Deployment
```python
model = AutoModelForCausalLM.from_pretrained(
"AlphaExaAI/ExaMind",
torch_dtype=torch.float32,
device_map="cpu"
)
```
### Using with llama.cpp (GGUF β€” Coming Soon)
```bash
# GGUF quantized versions will be released for efficient CPU inference
# Stay tuned for Q4_K_M, Q5_K_M, and Q8_0 variants
```
---
## πŸ—οΈ Architecture
```
ExaMind-V2-Final
β”œβ”€β”€ Architecture: Qwen2ForCausalLM (Transformer)
β”œβ”€β”€ Hidden Size: 3,584
β”œβ”€β”€ Intermediate Size: 18,944
β”œβ”€β”€ Layers: 28
β”œβ”€β”€ Attention Heads: 28
β”œβ”€β”€ KV Heads: 4 (GQA)
β”œβ”€β”€ Vocab Size: 152,064
β”œβ”€β”€ Max Position: 32,768 (extendable to 128K)
β”œβ”€β”€ Activation: SiLU
β”œβ”€β”€ RoPE ΞΈ: 1,000,000
└── Precision: FP32 / FP16 compatible
```
---
## πŸ› οΈ Training Methodology
ExaMind was developed using a multi-stage training pipeline:
| Stage | Method | Description |
|-------|--------|-------------|
| **Stage 1** | Base Model Selection | Qwen2.5-Coder-7B as foundation |
| **Stage 2** | Supervised Fine-Tuning (SFT) | Training on curated 2026 datasets |
| **Stage 3** | LoRA Adaptation | Low-Rank Adaptation for efficient specialization |
| **Stage 4** | Identity Enforcement | Hardcoded identity alignment and security tuning |
| **Stage 5** | Security Alignment | Prompt injection resistance training |
| **Stage 6** | Chat Template Integration | Custom Jinja2 template with system prompt |
---
## πŸ“š Training Data
### Public Data Sources
- Programming and code corpora (GitHub, StackOverflow)
- General web text and knowledge bases
- Technical documentation and research papers
- Multilingual text data
### Custom Alignment Data
- Identity enforcement instruction dataset
- Security-focused instruction tuning samples
- Prompt injection resistance adversarial examples
- Structured conversational datasets
- Complex problem-solving chains
> ⚠️ No private user data was used in training. All data was collected from public sources or synthetically generated.
---
## πŸ”’ Security Features
ExaMind includes built-in security measures:
- **Identity Lock** β€” The model maintains its ExaMind identity and cannot be tricked into impersonating other models
- **Prompt Injection Resistance** β€” 92% resistance rate against instruction override attacks
- **System Prompt Protection** β€” Refuses to reveal internal configuration or system prompts
- **Safe Output Generation** β€” Prioritizes safety and secure development practices
- **Hallucination Reduction** β€” States assumptions and avoids fabricating information
---
## πŸ“‹ Model Files
| File | Size | Description |
|------|------|-------------|
| `model.safetensors` | ~29 GB | Model weights (FP32) |
| `config.json` | 1.4 KB | Model configuration |
| `tokenizer.json` | 11 MB | Tokenizer vocabulary |
| `tokenizer_config.json` | 663 B | Tokenizer settings |
| `generation_config.json` | 241 B | Default generation parameters |
| `chat_template.jinja` | 1.4 KB | Chat template with system prompt |
---
## πŸ—ΊοΈ Roadmap
- [x] ExaMind V1 β€” Initial release
- [x] ExaMind V2-Final β€” Production-ready with security alignment
- [ ] ExaMind V2-GGUF β€” Quantized versions for CPU inference
- [ ] ExaMind V3 β€” Extended context (128K), improved reasoning
- [ ] ExaMind-Code β€” Specialized coding variant
- [ ] ExaMind-Vision β€” Multimodal capabilities
---
## 🀝 Contributing
We welcome contributions from the community! ExaMind is fully open-source and we're excited to collaborate.
### How to Contribute
1. **Fork** the repository on [GitHub](https://github.com/hleliofficiel/AlphaExaAI)
2. **Create** a feature branch (`git checkout -b feature/amazing-feature`)
3. **Commit** your changes (`git commit -m 'Add amazing feature'`)
4. **Push** to the branch (`git push origin feature/amazing-feature`)
5. **Open** a Pull Request
### Areas We Need Help
- πŸ§ͺ Benchmark evaluation on additional datasets
- 🌍 Multilingual evaluation and improvement
- πŸ“ Documentation and tutorials
- πŸ”§ Quantization and optimization
- πŸ›‘οΈ Security testing and red-teaming
---
## πŸ“„ License
This project is licensed under the **Apache License 2.0** β€” see the [LICENSE](LICENSE) file for details.
You are free to:
- βœ… Use commercially
- βœ… Modify and distribute
- βœ… Use privately
- βœ… Patent use
---
## πŸ“¬ Contact
- **Organization:** [AlphaExaAI](https://huggingface.co/AlphaExaAI)
- **GitHub:** [github.com/hleliofficiel/AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI)
- **Email:** h.hleli@tuta.io
---
<div align="center">
**Built with ❀️ by AlphaExaAI Team β€” 2026**
*Advancing open-source AI, one model at a time.*
</div>