|
|
--- |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
- ar |
|
|
- fr |
|
|
- zh |
|
|
- de |
|
|
- es |
|
|
- ja |
|
|
- ko |
|
|
- ru |
|
|
- pt |
|
|
- multilingual |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- qwen2 |
|
|
- chat |
|
|
- code |
|
|
- security |
|
|
- alphaexaai |
|
|
- examind |
|
|
- conversational |
|
|
- open-source |
|
|
base_model: |
|
|
- Qwen/Qwen2.5-Coder-7B |
|
|
model-index: |
|
|
- name: ExaMind-V2-Final |
|
|
results: |
|
|
- task: |
|
|
type: text-generation |
|
|
name: Text Generation |
|
|
dataset: |
|
|
name: MMLU |
|
|
type: cais/mmlu |
|
|
metrics: |
|
|
- type: accuracy |
|
|
name: MMLU World Religions (0-shot) |
|
|
value: 94.8 |
|
|
verified: false |
|
|
- task: |
|
|
type: text-generation |
|
|
name: Code Generation |
|
|
dataset: |
|
|
name: HumanEval |
|
|
type: openai/openai_humaneval |
|
|
metrics: |
|
|
- type: pass@1 |
|
|
name: HumanEval pass@1 |
|
|
value: 79.3 |
|
|
verified: false |
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
# π§ ExaMind |
|
|
|
|
|
### Advanced Open-Source AI by AlphaExaAI |
|
|
|
|
|
[](https://opensource.org/licenses/Apache-2.0) |
|
|
[](https://huggingface.co/AlphaExaAI/ExaMind) |
|
|
[](https://github.com/hleliofficiel/AlphaExaAI) |
|
|
[](https://huggingface.co/Qwen) |
|
|
|
|
|
**ExaMind** is an advanced open-source conversational AI model developed by the **AlphaExaAI** team. |
|
|
Designed for secure, structured, and professional AI assistance with strong identity enforcement and production-ready deployment stability. |
|
|
|
|
|
[π Get Started](#-quick-start) Β· [π Benchmarks](#-benchmarks) Β· [π€ Contributing](#-contributing) Β· [π License](#-license) |
|
|
|
|
|
</div> |
|
|
|
|
|
--- |
|
|
|
|
|
## π Model Overview |
|
|
|
|
|
| Property | Details | |
|
|
|----------|---------| |
|
|
| **Model Name** | ExaMind | |
|
|
| **Version** | V2-Final | |
|
|
| **Developer** | [AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI) | |
|
|
| **Base Architecture** | Qwen2.5-Coder-7B | |
|
|
| **Parameters** | 7.62 Billion (~8B) | |
|
|
| **Precision** | FP32 (~29GB) / FP16 (~15GB) | |
|
|
| **Context Window** | 32,768 tokens (supports up to 128K with RoPE scaling) | |
|
|
| **License** | Apache 2.0 | |
|
|
| **Languages** | Multilingual (English preferred) | |
|
|
| **Deployment** | β
CPU & GPU compatible | |
|
|
|
|
|
--- |
|
|
|
|
|
## β¨ Key Capabilities |
|
|
|
|
|
- π₯οΈ **Advanced Programming** β Code generation, debugging, architecture design, and code review |
|
|
- π§© **Complex Problem Solving** β Multi-step logical reasoning and deep technical analysis |
|
|
- π **Security-First Design** β Built-in prompt injection resistance and identity enforcement |
|
|
- π **Multilingual** β Supports all major world languages, optimized for English |
|
|
- π¬ **Conversational AI** β Natural, structured, and professional dialogue |
|
|
- ποΈ **Scalable Architecture** β Secure software engineering and system design guidance |
|
|
- β‘ **CPU Deployable** β Runs on CPU nodes without GPU requirement |
|
|
|
|
|
--- |
|
|
|
|
|
## π Benchmarks |
|
|
|
|
|
### General Knowledge & Reasoning |
|
|
|
|
|
| Benchmark | Setting | Score | |
|
|
|-----------|---------|-------| |
|
|
| **MMLU β World Religions** | 0-shot | **94.8%** | |
|
|
| **MMLU β Overall** | 5-shot | **72.1%** | |
|
|
| **ARC-Challenge** | 25-shot | **68.4%** | |
|
|
| **HellaSwag** | 10-shot | **78.9%** | |
|
|
| **TruthfulQA** | 0-shot | **61.2%** | |
|
|
| **Winogrande** | 5-shot | **74.5%** | |
|
|
|
|
|
### Code Generation |
|
|
|
|
|
| Benchmark | Setting | Score | |
|
|
|-----------|---------|-------| |
|
|
| **HumanEval** | pass@1 | **79.3%** | |
|
|
| **MBPP** | pass@1 | **71.8%** | |
|
|
| **MultiPL-E (Python)** | pass@1 | **76.5%** | |
|
|
| **DS-1000** | pass@1 | **48.2%** | |
|
|
|
|
|
### Math & Reasoning |
|
|
|
|
|
| Benchmark | Setting | Score | |
|
|
|-----------|---------|-------| |
|
|
| **GSM8K** | 8-shot CoT | **82.4%** | |
|
|
| **MATH** | 4-shot | **45.7%** | |
|
|
|
|
|
### π Prompt Injection Resistance |
|
|
|
|
|
| Test | Details | |
|
|
|------|---------| |
|
|
| **Test Set Size** | 50 adversarial prompts | |
|
|
| **Attack Type** | Instruction override / identity manipulation | |
|
|
| **Resistance Rate** | **92%** | |
|
|
| **Method** | Custom red-teaming with jailbreak & override attempts | |
|
|
|
|
|
> Evaluation performed using `lm-eval-harness` on CPU. Security tests performed using custom adversarial prompt suite. |
|
|
|
|
|
--- |
|
|
|
|
|
## π Quick Start |
|
|
|
|
|
### Installation |
|
|
|
|
|
```bash |
|
|
pip install transformers torch accelerate |
|
|
``` |
|
|
|
|
|
### Basic Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
import torch |
|
|
|
|
|
model_path = "AlphaExaAI/ExaMind" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(model_path) |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
model_path, |
|
|
torch_dtype=torch.float16, |
|
|
device_map="auto" |
|
|
) |
|
|
|
|
|
messages = [ |
|
|
{"role": "user", "content": "Explain how to secure a REST API."} |
|
|
] |
|
|
|
|
|
inputs = tokenizer.apply_chat_template( |
|
|
messages, |
|
|
return_tensors="pt", |
|
|
add_generation_prompt=True |
|
|
).to(model.device) |
|
|
|
|
|
outputs = model.generate( |
|
|
inputs, |
|
|
max_new_tokens=512, |
|
|
temperature=0.7, |
|
|
top_p=0.8, |
|
|
top_k=20, |
|
|
repetition_penalty=1.1 |
|
|
) |
|
|
|
|
|
response = tokenizer.decode( |
|
|
outputs[0][inputs.shape[-1]:], |
|
|
skip_special_tokens=True |
|
|
) |
|
|
print(response) |
|
|
``` |
|
|
|
|
|
### CPU Deployment |
|
|
|
|
|
```python |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
"AlphaExaAI/ExaMind", |
|
|
torch_dtype=torch.float32, |
|
|
device_map="cpu" |
|
|
) |
|
|
``` |
|
|
|
|
|
### Using with llama.cpp (GGUF β Coming Soon) |
|
|
|
|
|
```bash |
|
|
# GGUF quantized versions will be released for efficient CPU inference |
|
|
# Stay tuned for Q4_K_M, Q5_K_M, and Q8_0 variants |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## ποΈ Architecture |
|
|
|
|
|
``` |
|
|
ExaMind-V2-Final |
|
|
βββ Architecture: Qwen2ForCausalLM (Transformer) |
|
|
βββ Hidden Size: 3,584 |
|
|
βββ Intermediate Size: 18,944 |
|
|
βββ Layers: 28 |
|
|
βββ Attention Heads: 28 |
|
|
βββ KV Heads: 4 (GQA) |
|
|
βββ Vocab Size: 152,064 |
|
|
βββ Max Position: 32,768 (extendable to 128K) |
|
|
βββ Activation: SiLU |
|
|
βββ RoPE ΞΈ: 1,000,000 |
|
|
βββ Precision: FP32 / FP16 compatible |
|
|
``` |
|
|
|
|
|
--- |
|
|
|
|
|
## π οΈ Training Methodology |
|
|
|
|
|
ExaMind was developed using a multi-stage training pipeline: |
|
|
|
|
|
| Stage | Method | Description | |
|
|
|-------|--------|-------------| |
|
|
| **Stage 1** | Base Model Selection | Qwen2.5-Coder-7B as foundation | |
|
|
| **Stage 2** | Supervised Fine-Tuning (SFT) | Training on curated 2026 datasets | |
|
|
| **Stage 3** | LoRA Adaptation | Low-Rank Adaptation for efficient specialization | |
|
|
| **Stage 4** | Identity Enforcement | Hardcoded identity alignment and security tuning | |
|
|
| **Stage 5** | Security Alignment | Prompt injection resistance training | |
|
|
| **Stage 6** | Chat Template Integration | Custom Jinja2 template with system prompt | |
|
|
|
|
|
--- |
|
|
|
|
|
## π Training Data |
|
|
|
|
|
### Public Data Sources |
|
|
- Programming and code corpora (GitHub, StackOverflow) |
|
|
- General web text and knowledge bases |
|
|
- Technical documentation and research papers |
|
|
- Multilingual text data |
|
|
|
|
|
### Custom Alignment Data |
|
|
- Identity enforcement instruction dataset |
|
|
- Security-focused instruction tuning samples |
|
|
- Prompt injection resistance adversarial examples |
|
|
- Structured conversational datasets |
|
|
- Complex problem-solving chains |
|
|
|
|
|
> β οΈ No private user data was used in training. All data was collected from public sources or synthetically generated. |
|
|
|
|
|
--- |
|
|
|
|
|
## π Security Features |
|
|
|
|
|
ExaMind includes built-in security measures: |
|
|
|
|
|
- **Identity Lock** β The model maintains its ExaMind identity and cannot be tricked into impersonating other models |
|
|
- **Prompt Injection Resistance** β 92% resistance rate against instruction override attacks |
|
|
- **System Prompt Protection** β Refuses to reveal internal configuration or system prompts |
|
|
- **Safe Output Generation** β Prioritizes safety and secure development practices |
|
|
- **Hallucination Reduction** β States assumptions and avoids fabricating information |
|
|
|
|
|
--- |
|
|
|
|
|
## π Model Files |
|
|
|
|
|
| File | Size | Description | |
|
|
|------|------|-------------| |
|
|
| `model.safetensors` | ~29 GB | Model weights (FP32) | |
|
|
| `config.json` | 1.4 KB | Model configuration | |
|
|
| `tokenizer.json` | 11 MB | Tokenizer vocabulary | |
|
|
| `tokenizer_config.json` | 663 B | Tokenizer settings | |
|
|
| `generation_config.json` | 241 B | Default generation parameters | |
|
|
| `chat_template.jinja` | 1.4 KB | Chat template with system prompt | |
|
|
|
|
|
--- |
|
|
|
|
|
## πΊοΈ Roadmap |
|
|
|
|
|
- [x] ExaMind V1 β Initial release |
|
|
- [x] ExaMind V2-Final β Production-ready with security alignment |
|
|
- [ ] ExaMind V2-GGUF β Quantized versions for CPU inference |
|
|
- [ ] ExaMind V3 β Extended context (128K), improved reasoning |
|
|
- [ ] ExaMind-Code β Specialized coding variant |
|
|
- [ ] ExaMind-Vision β Multimodal capabilities |
|
|
|
|
|
--- |
|
|
|
|
|
## π€ Contributing |
|
|
|
|
|
We welcome contributions from the community! ExaMind is fully open-source and we're excited to collaborate. |
|
|
|
|
|
### How to Contribute |
|
|
|
|
|
1. **Fork** the repository on [GitHub](https://github.com/hleliofficiel/AlphaExaAI) |
|
|
2. **Create** a feature branch (`git checkout -b feature/amazing-feature`) |
|
|
3. **Commit** your changes (`git commit -m 'Add amazing feature'`) |
|
|
4. **Push** to the branch (`git push origin feature/amazing-feature`) |
|
|
5. **Open** a Pull Request |
|
|
|
|
|
### Areas We Need Help |
|
|
|
|
|
- π§ͺ Benchmark evaluation on additional datasets |
|
|
- π Multilingual evaluation and improvement |
|
|
- π Documentation and tutorials |
|
|
- π§ Quantization and optimization |
|
|
- π‘οΈ Security testing and red-teaming |
|
|
|
|
|
--- |
|
|
|
|
|
## π License |
|
|
|
|
|
This project is licensed under the **Apache License 2.0** β see the [LICENSE](LICENSE) file for details. |
|
|
|
|
|
You are free to: |
|
|
- β
Use commercially |
|
|
- β
Modify and distribute |
|
|
- β
Use privately |
|
|
- β
Patent use |
|
|
|
|
|
--- |
|
|
|
|
|
## π¬ Contact |
|
|
|
|
|
- **Organization:** [AlphaExaAI](https://huggingface.co/AlphaExaAI) |
|
|
- **GitHub:** [github.com/hleliofficiel/AlphaExaAI](https://github.com/hleliofficiel/AlphaExaAI) |
|
|
- **Email:** h.hleli@tuta.io |
|
|
|
|
|
--- |
|
|
|
|
|
<div align="center"> |
|
|
|
|
|
**Built with β€οΈ by AlphaExaAI Team β 2026** |
|
|
|
|
|
*Advancing open-source AI, one model at a time.* |
|
|
|
|
|
</div> |
|
|
|