File size: 7,833 Bytes
7c5aef2
73fb0f9
 
 
7c5aef2
 
 
73fb0f9
 
 
 
7c5aef2
 
73fb0f9
 
 
 
 
 
 
 
 
 
 
 
 
74605de
73fb0f9
 
74605de
 
73fb0f9
7c5aef2
 
73fb0f9
 
 
 
 
7c5aef2
74605de
7c5aef2
73fb0f9
 
 
 
 
 
 
 
 
 
 
 
 
7c5aef2
21e4a83
74605de
7c5aef2
74605de
21e4a83
73fb0f9
 
74605de
21e4a83
74605de
7c5aef2
73fb0f9
7c5aef2
74605de
7c5aef2
74605de
73fb0f9
 
7c5aef2
74605de
21e4a83
73fb0f9
 
 
 
 
21e4a83
74605de
7c5aef2
73fb0f9
 
7c5aef2
73fb0f9
21e4a83
74605de
73fb0f9
74605de
21e4a83
74605de
 
 
 
 
 
73fb0f9
74605de
73fb0f9
74605de
 
 
 
 
 
 
 
 
73fb0f9
 
 
 
 
 
 
 
 
 
 
 
74605de
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
73fb0f9
74605de
73fb0f9
 
 
74605de
73fb0f9
74605de
73fb0f9
 
7c5aef2
 
73fb0f9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7c5aef2
 
73fb0f9
 
 
7c5aef2
74605de
7c5aef2
73fb0f9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7c5aef2
 
74605de
7c5aef2
73fb0f9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7c5aef2
 
74605de
7c5aef2
 
 
 
73fb0f9
7c5aef2
 
73fb0f9
 
 
 
 
21e4a83
74605de
21e4a83
73fb0f9
7c5aef2
74605de
7c5aef2
73fb0f9
 
7c5aef2
73fb0f9
 
74605de
73fb0f9
 
 
74605de
7c5aef2
73fb0f9
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
---
library_name: transformers
license: mit
pipeline_tag: text-generation
language:
- en
- code
tags:
- transformers
- pytorch
- safetensors
- text-generation
- code-generation
- python
- javascript
- coding
- programming
- sagemaker
- amazon-sagemaker
- cpu
- compact
- efficient
- nvdya-kit
- death-legion
- vllm
- sglang
- llama-cpp
- ollama
- lm-studio
- year-2026
- next-gen

datasets:
- the-stack-v2

metrics:
- perplexity
- accuracy

model-index:
- name: Legion Coder 8M 2026
  results: []

inference:
  parameters:
    temperature: 0.8
    top_p: 0.95
    top_k: 50
    max_new_tokens: 200

sagemaker:
  sdk_version: "2.200.0"
  instance_type: "ml.m5.large"
  instance_count: 1
  container_image: "huggingface-pytorch-inference:2.0.0-transformers4.28.1-cpu-py310-ubuntu20.04-v1.0"
---

# Legion Coder 8M 2026

**A 44M Parameter Transformer for Code Generation - 2026 Edition**

[![Made with by DEATH LEGION](https://img.shields.io/badge/MADE%20WITH%20BY-DEATH%20LEGION-ff0040?style=for-the-badge)](https://huggingface.co/dineth554/legion-coder-8m)
[![Powered by nvdya-kit](https://img.shields.io/badge/POWERED%20BY-nvdya--kit-7c4dff?style=for-the-badge)]()
[![2026 Edition](https://img.shields.io/badge/2026-EDITION-00d4ff?style=for-the-badge)]()

## Quick Links

<div align="center">

### Libraries and Frameworks

[![Transformers](https://img.shields.io/badge/Transformers-Compatible-brightgreen?style=flat-square&logo=huggingface)](https://huggingface.co/docs/transformers)
[![PyTorch](https://img.shields.io/badge/PyTorch-2.1+-ee4c2c?style=flat-square&logo=pytorch)](https://pytorch.org/)
[![Safetensors](https://img.shields.io/badge/Safetensors-Format-blue?style=flat-square)](https://github.com/huggingface/safetensors)

### Local Apps and Inference Engines

[![vLLM](https://img.shields.io/badge/vLLM-Supported-ff6b6b?style=flat-square)](https://docs.vllm.ai/)
[![SGLang](https://img.shields.io/badge/SGLang-New!-4ecdc4?style=flat-square)](https://sgl-project.github.io/)
[![llama.cpp](https://img.shields.io/badge/llama.cpp-Compatible-8b5cf6?style=flat-square)](https://github.com/ggerganov/llama.cpp)
[![Ollama](https://img.shields.io/badge/Ollama-Ready-f97316?style=flat-square)](https://ollama.ai/)
[![LM Studio](https://img.shields.io/badge/LM%20Studio-Compatible-10b981?style=flat-square)](https://lmstudio.ai/)

### Notebooks and Cloud

[![Open In Colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/github/dineth554/legion-coder-8m/blob/main/notebooks/legion_coder_demo.ipynb)
[![Kaggle](https://kaggle.com/static/images/open-in-kaggle.svg)](https://kaggle.com/kernels/welcome?src=https://github.com/dineth554/legion-coder-8m/blob/main/notebooks/legion_coder_demo.ipynb)

</div>

## About

Legion Coder 2026 is a compact yet powerful 44M parameter transformer model optimized for coding tasks. Built with precision by **DEATH LEGION** and powered by **nvdya-kit**, this model delivers high-quality code generation in a lightweight package.

**2026 Edition Features:**
- Enhanced performance optimizations
- Updated documentation and branding
- Professional icon-based UI
- Advanced CSS animations
- Performance comparison charts

## Features

- **Clean Code Generation** - PEP 8 compliant Python and more
- **Debug Assistance** - Help identify and fix code issues
- **Code Explanation** - Understand complex programming concepts
- **Multi-language Support** - Python, JavaScript, and more
- **Fast Inference** - Optimized for CPU deployment
- **SageMaker Ready** - One-click AWS deployment
- **Template Ready** - Duplicate this space to create your own

## Model Specifications 2026

| Attribute | Value |
|-----------|-------|
| **Parameters** | 44,341,632 (~44M) |
| **Model Size** | ~170MB |
| **Architecture** | GPT-style Transformer |
| **Hidden Size** | 576 |
| **Layers** | 13 |
| **Attention Heads** | 16 |
| **Context Length** | 1,024 tokens |
| **Vocabulary** | 16,000 tokens |
| **Format** | Safetensors |
| **Edition** | 2026 |

## Model Comparison 2026

| Model | Parameters | Size | Efficiency Score | Best For |
|-------|------------|------|------------------|----------|
| **Legion Coder 8M** | 44M | ~170MB | 9.5/10 | Code generation, CPU inference |
| TinyLlama-1.1B | 1.1B | ~2.2GB | 6.0/10 | General text, GPU required |
| Qwen2.5-0.5B | 500M | ~1.0GB | 7.0/10 | Multilingual, GPU recommended |
| CodeLlama-7B | 7B | ~13GB | 5.0/10 | Production code, GPU required |
| Phi-2 | 2.7B | ~5.3GB | 6.5/10 | Reasoning, GPU required |

**Efficiency Score** = (Parameter Efficiency x Memory Efficiency x Speed) / 3

Legion Coder 8M 2026 achieves exceptional efficiency through:
- **260x smaller** than CodeLlama-7B
- **13x smaller** than TinyLlama-1.1B
- **6x smaller** than Qwen2.5-0.5B
- Runs entirely on CPU with 8GB RAM

## Amazon SageMaker Deployment

This model is ready for deployment on Amazon SageMaker with one-click deployment support.

### Deploy to AWS SageMaker

[![Deploy to SageMaker](https://img.shields.io/badge/Deploy%20to-AWS%20SageMaker-FF9900?style=for-the-badge&logo=amazon-aws)](https://huggingface.co/dineth554/legion-coder-8m/deploy/sagemaker)

### Using the SageMaker Python SDK

```python
import sagemaker
from sagemaker.huggingface import HuggingFaceModel

# Initialize SageMaker session
sess = sagemaker.Session()

# Create Hugging Face Model
huggingface_model = HuggingFaceModel(
    model_data="dineth554/legion-coder-8m",
    transformers_version="4.36.0",
    pytorch_version="2.1.0",
    py_version="py310",
    role="arn:aws:iam::YOUR_ACCOUNT_ID:role/YOUR_SAGEMAKER_ROLE",
    sagemaker_session=sess,
)

# Deploy to SageMaker
predictor = huggingface_model.deploy(
    initial_instance_count=1,
    instance_type="ml.m5.large",
    endpoint_name="legion-coder-8m-endpoint"
)

# Test the endpoint
result = predictor.predict({
    "inputs": "Write a Python function to calculate fibonacci numbers:",
    "parameters": {
        "temperature": 0.8,
        "max_new_tokens": 200
    }
})

print(result)
```

### SageMaker Inference Script

The `sagemaker_inference.py` file in this repository provides the inference handler for SageMaker deployment.

## Local Inference with vLLM

```python
from vllm import LLM, SamplingParams

# Load model with vLLM
llm = LLM(model="dineth554/legion-coder-8m")

# Set sampling parameters
sampling_params = SamplingParams(
    temperature=0.8,
    top_p=0.95,
    max_tokens=200
)

# Generate code
prompt = "Write a Python function to calculate fibonacci numbers:"
outputs = llm.generate(prompt, sampling_params)
print(outputs[0].outputs[0].text)
```

## Local Inference with SGLang

```python
import sglang as sgl

# Define prompt template
@sgl.function
def code_gen(s, prompt):
    s += sgl.system("You are a helpful coding assistant.")
    s += sgl.user(prompt)
    s += sgl.assistant(sgl.gen("code", max_tokens=200))

# Run inference
result = code_gen.run(
    prompt="Write a Python function to calculate fibonacci numbers:",
    temperature=0.8
)
print(result["code"])
```

## Technical Details

### Training Data
- Python code from The Stack v2 dataset
- GitHub code repositories (filtered for quality)
- Code-specific preprocessing for indentation and special tokens

### Training Procedure
- **Optimizer:** AdamW
- **Learning Rate:** 5e-4 with cosine decay
- **Batch Size:** 4 with gradient accumulation
- **Training Steps:** 10,000
- **Precision:** float32 (CPU-optimized)

## License

This model is released under the **MIT License**.

## Links

- **Model Repository:** [dineth554/legion-coder-8m](https://huggingface.co/dineth554/legion-coder-8m)
- **Live Demo:** [Hugging Face Space](https://huggingface.co/spaces/dineth554/legion-coder-8m)

<div align="center">

### MADE WITH BY DEATH LEGION

**Powered by nvdya-kit**

*2026 DEATH LEGION. All rights reserved.*

</div>