File size: 3,702 Bytes
6649032
 
 
 
 
eff4789
 
6649032
 
 
5d90d09
6649032
 
 
 
5d90d09
6649032
11320a7
 
6649032
11320a7
6649032
 
 
11320a7
6649032
5d90d09
6649032
11320a7
6649032
11320a7
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
065e778
5d90d09
11320a7
 
b21abe1
11320a7
 
 
2530cc5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11320a7
 
 
 
bf9e57e
11320a7
 
 
 
 
 
 
 
 
 
6649032
11320a7
6649032
5d90d09
6649032
 
 
11320a7
6649032
11320a7
2530cc5
11320a7
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
---
license: mit
language:
- en
tags:
- student-startup
- zero-to-one
- nef
- solo-developer
- bangladesh-ai
- 2b-parameters
pipeline_tag: text-generation
library_name: pytorch
---

# Hexa-2B — NEF Serialization Prototype

**Founder:** Madhab — Engineering Student, Cox's Bazar, Bangladesh  
**Organization:** Hexa Innovate  
**Format:** [NEF (Neural Essence Format)](https://github.com/Hexa08/NEF)  
**Purpose:** Infrastructure validation prototype — not a production inference model

---

## What This Is

Hexa-2B is a 2-billion parameter language model built as a **technical proof-of-concept for the NEF serialization framework**. The goal of this release is singular: demonstrate that NEF can correctly serialize, store, and load a billion-scale model on accessible hardware without dependency on standard bloated AI libraries.

This is not a general-purpose chat model. Inference quality is intentionally deferred to the production training run. What this prototype proves is the infrastructure layer — and that is the point.

---

## NEF — Neural Essence Format

NEF is a custom serialization framework built from scratch to replace the overhead of standard formats (safetensors, GGUF, Pickle) for open-weight model loading.

| Property | Detail |
|---|---|
| Layout | Flat binary, memory-mapped tensor access |
| Runtime deps | None |
| Target | Fast loading on mid-range and edge hardware |
| Status | Active development |

**Repository:** [github.com/Hexa08/NEF](https://github.com/Hexa08/NEF)

---

## Technical Specs

| Property | Detail |
|---|---|
| Architecture | Mixture OF Expart |
| Parameters | 2 Billion (0.27B active via MoE) |
| Serialization | NEF (Neural Essence Format) |
| Training hardware | Dual NVIDIA Tesla T4 (cloud compute credits) |
| Languages | English |

---

## Benchmark Results

Early checkpoint evaluation (step 40,000) on standard zero-shot benchmarks against GPT-2 124M baseline:

![Benchmark Results](assets/benchmark.png)

| Task | Hexa 2B (MoE) | GPT-2 124M | Delta |
|---|---|---|---|
| ARC Easy | 26.5% | 43.2% | -16.7% |
| ARC Challenge | **27.0%** | 22.4% | **+4.6%** |
| OpenBookQA | **25.0%** | 14.2% | **+10.8%** |
| WinoGrande | 47.9% | 51.3% | -3.4% |
| **Average** | **31.6%** | 32.8% | -1.2% |

> Zero-shot evaluation using [EleutherAI lm-evaluation-harness](https://github.com/EleutherAI/lm-evaluation-harness) v0.4.2 at training step 40,000. 2 out of 4 tasks already exceed GPT-2 124M. Full evaluation pending production training run.

---

## Prototype Scope

This release validates the following:

- NEF correctly serializes 2.1B parameters to disk
- NEF correctly deserializes and loads the full model into memory
- The full pipeline runs on accessible hardware without enterprise infrastructure

**Inference benchmarks and model quality evaluations are reserved for the next training run**, which uses a larger, high-diversity multilingual corpus and a production-grade training configuration.

---

## Founder

I am a Diploma in Engineering student from Cox's Bazar, Bangladesh. Every component of this project — the HexaDense architecture, the NEF serialization format, and the training pipeline — was engineered solo, with no external funding and no institutional backing.

Most billion-parameter models come from large teams with large budgets. This one did not. The constraint was the design brief.

Hexa-2B is the foundation. The production model is next.

---

## About Hexa Innovate

Hexa Innovate is a student-led AI startup based in Bangladesh, focused on building efficient AI execution and serialization infrastructure for open-weight models at the edge.

**GitHub:** [github.com/Hexa08](https://github.com/Hexa08)