legion-coder-8m / README.yaml
dineth554's picture
Upload README.yaml with huggingface_hub
b106aa6 verified
# Model Card for Legion Coder 8M
# YAML Front Matter for Hugging Face Hub
base_model: dineth554/legion-coder-8m
library_name: transformers
license: mit
pipeline_tag: text-generation
language:
- en
- code
tags:
- transformers
- pytorch
- safetensors
- text-generation
- code-generation
- python
- javascript
- coding
- programming
- sagemaker
- amazon-sagemaker
- cpu
- compact
- efficient
- nvdya-kit
- death-legion
datasets:
- the-stack-v2
metrics:
- perplexity
- accuracy
model-index:
- name: Legion Coder 8M
results: []
inference:
parameters:
temperature: 0.8
top_p: 0.95
top_k: 50
max_new_tokens: 200
sagemaker:
sdk_version: "2.200.0"
instance_type: "ml.m5.large"
instance_count: 1
container_image: "huggingface-pytorch-inference:2.0.0-transformers4.28.1-cpu-py310-ubuntu20.04-v1.0"
# Model Details
model_details:
name: Legion Coder 8M
version: 1.0.0
description: A compact yet powerful 44M parameter transformer model optimized for coding tasks
developer: DEATH LEGION
powered_by: nvdya-kit
architecture: GPT-style Transformer
parameters: 44,341,632
model_size: 170MB
hidden_size: 576
num_layers: 13
num_heads: 16
context_length: 1024
vocabulary_size: 16000
format: Safetensors
precision: float32
# Training Details
training_details:
optimizer: AdamW
learning_rate: 5e-4
lr_schedule: cosine_decay
batch_size: 4
gradient_accumulation: true
training_steps: 10000
precision: float32
# Intended Use
intended_use:
primary_use_cases:
- Code completion and generation
- Function generation from descriptions
- Debugging assistance
- Code explanation and documentation
- Programming concept explanations
- Code scaffolding and prototyping
target_users:
- Software developers
- Students learning to code
- Data scientists
- DevOps engineers
- Technical writers
# Limitations
limitations:
- Limited to 1,024 token context window
- Trained primarily on Python code
- May generate code that requires review before production use
- Not suitable for non-coding tasks
# Ethical Considerations
ethical_considerations:
- Generated code should be reviewed before deployment
- May reproduce patterns from training data
- Not a replacement for human code review
- Users are responsible for compliance with licenses of generated code
# Citation
citation: |
@misc{legioncoder2026,
title={Legion Coder 8M: A Compact Transformer for Code Generation},
author={DEATH LEGION},
year={2026},
howpublished={\url{https://huggingface.co/dineth554/legion-coder-8m}}
}
# Contact
contact:
developer: DEATH LEGION
powered_by: nvdya-kit
repository: https://huggingface.co/dineth554/legion-coder-8m
# Branding
branding:
tagline: MADE WITH BY DEATH LEGION
powered_by: nvdya-kit
copyright: 2026 DEATH LEGION. All rights reserved.