|
|
--- |
|
|
base_model: codellama/CodeLlama-7b-Instruct-hf |
|
|
library_name: peft |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- base_model:adapter:codellama/CodeLlama-7b-Instruct-hf |
|
|
- lora |
|
|
- transformers |
|
|
- luau |
|
|
- roblox |
|
|
license: apache-2.0 |
|
|
language: |
|
|
- en |
|
|
--- |
|
|
|
|
|
# Model Card for CodeLlama-7B-Instruct-Luau |
|
|
|
|
|
Fine-tuned version of `codellama/CodeLlama-7b-Instruct-hf` targeted toward the **Luau** programming language, Roblox’s Lua-derived scripting language. |
|
|
|
|
|
This model is distributed as a **LoRA adapter** and is intended to improve the base model’s performance on Roblox-specific scripting tasks. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Details |
|
|
|
|
|
### Model Description |
|
|
|
|
|
This model is a parameter-efficient fine-tuning (LoRA) of CodeLlama 7B Instruct, specialized for generating, explaining, and refactoring **Luau** code. |
|
|
|
|
|
The fine-tuning focuses on Roblox development patterns, including common services, APIs, gameplay scripting idioms, and client/server logic. The model is designed to assist developers during prototyping, learning, and general scripting workflows. |
|
|
|
|
|
- **Developed by:** darwinkernelpanic |
|
|
- **Funded by:** Not applicable |
|
|
- **Shared by:** darwinkernelpanic |
|
|
- **Model type:** Causal Language Model (decoder-only, LoRA adapter) |
|
|
- **Language(s) (NLP):** English |
|
|
- **License:** Apache-2.0 |
|
|
- **Finetuned from model:** codellama/CodeLlama-7b-Instruct-hf |
|
|
|
|
|
### Model Sources |
|
|
|
|
|
- **Repository:** https://huggingface.co/darwinkernelpanic/CodeLlama-7b-Instruct-hf-luau |
|
|
- **Paper:** *Code Llama: Large Language Models for Code* (Meta AI) |
|
|
- **Demo:** Not available |
|
|
|
|
|
--- |
|
|
|
|
|
## Uses |
|
|
|
|
|
### Direct Use |
|
|
|
|
|
This model can be used directly for: |
|
|
|
|
|
- Writing Luau scripts for Roblox |
|
|
- Explaining Roblox APIs and services |
|
|
- Refactoring or debugging Luau code |
|
|
- Prototyping gameplay systems and utilities |
|
|
- Learning Luau and Roblox scripting concepts |
|
|
|
|
|
The model is intended as a **developer assistant**, not an autonomous system. |
|
|
|
|
|
### Downstream Use |
|
|
|
|
|
Potential downstream uses include: |
|
|
|
|
|
- Further fine-tuning on proprietary Roblox frameworks |
|
|
- Integration into IDEs or editor tooling |
|
|
- Chat-based assistants for Roblox development |
|
|
- Educational or documentation tooling |
|
|
|
|
|
### Out-of-Scope Use |
|
|
|
|
|
This model should **not** be used for: |
|
|
|
|
|
- Safety-critical or production-critical systems |
|
|
- Legal, medical, or financial advice |
|
|
- Malware, exploit, or cheat development |
|
|
- Fully automated code deployment without review |
|
|
|
|
|
--- |
|
|
|
|
|
## Bias, Risks, and Limitations |
|
|
|
|
|
- Inherits biases and limitations from the base CodeLlama model |
|
|
- May hallucinate Roblox APIs or outdated behaviors |
|
|
- Does not validate code at runtime |
|
|
- Output correctness depends on prompt quality |
|
|
|
|
|
### Recommendations |
|
|
|
|
|
Users should: |
|
|
|
|
|
- Review all generated code manually |
|
|
- Test scripts in Roblox Studio |
|
|
- Cross-check with official Roblox documentation |
|
|
- Treat outputs as suggestions rather than authoritative solutions |
|
|
|
|
|
--- |
|
|
|
|
|
## How to Get Started with the Model |
|
|
|
|
|
```python |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
from peft import PeftModel |
|
|
|
|
|
base_model = "codellama/CodeLlama-7b-Instruct-hf" |
|
|
adapter_model = "darwinkernelpanic/CodeLlama-7b-Instruct-hf-luau" |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained(base_model) |
|
|
model = AutoModelForCausalLM.from_pretrained(base_model) |
|
|
model = PeftModel.from_pretrained(model, adapter_model) |
|
|
|
|
|
prompt = "Write a Luau function that creates a Part and parents it to Workspace." |
|
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
|
|
|
|
outputs = model.generate( |
|
|
**inputs, |
|
|
max_new_tokens=300, |
|
|
temperature=0.7, |
|
|
do_sample=True |
|
|
) |
|
|
|
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
|
```` |
|
|
|
|
|
--- |
|
|
|
|
|
## Training Details |
|
|
|
|
|
### Training Data |
|
|
|
|
|
The model was fine-tuned on a curated mixture of: |
|
|
|
|
|
* Luau scripts |
|
|
* Roblox API usage examples |
|
|
* Open-source Roblox projects |
|
|
* Synthetic instruction-style prompts |
|
|
|
|
|
All data was filtered to avoid private, proprietary, or sensitive content. |
|
|
|
|
|
### Training Procedure |
|
|
|
|
|
The model was trained using parameter-efficient fine-tuning with LoRA while keeping the base model weights frozen. |
|
|
|
|
|
#### Preprocessing |
|
|
|
|
|
* Code formatting normalization |
|
|
* Instruction-style prompt structuring |
|
|
* Removal of low-quality or irrelevant samples |
|
|
|
|
|
#### Training Hyperparameters |
|
|
|
|
|
* **Training regime:** fp16 mixed precision |
|
|
|
|
|
#### Speeds, Sizes, Times |
|
|
|
|
|
* **Base model size:** ~7B parameters |
|
|
* **Trainable parameters:** <1% (LoRA adapters only) |
|
|
* **Adapter checkpoint size:** ~100–200 MB |
|
|
|
|
|
--- |
|
|
|
|
|
## Evaluation |
|
|
|
|
|
### Testing Data, Factors & Metrics |
|
|
|
|
|
#### Testing Data |
|
|
|
|
|
* Hand-written Luau prompts |
|
|
* Roblox-specific scripting scenarios |
|
|
|
|
|
#### Factors |
|
|
|
|
|
* Luau syntax correctness |
|
|
* Roblox API familiarity |
|
|
* Instruction-following behavior |
|
|
|
|
|
#### Metrics |
|
|
|
|
|
* Qualitative human evaluation |
|
|
* Manual code review and comparison with base model |
|
|
|
|
|
### Results |
|
|
|
|
|
The LoRA adapter demonstrates improved performance over the base model in: |
|
|
|
|
|
* Generating idiomatic Luau |
|
|
* Correct Roblox service usage |
|
|
* Following game-development-oriented instructions |
|
|
|
|
|
#### Summary |
|
|
|
|
|
The model performs best when used as a Roblox development assistant and is not intended for general-purpose natural language tasks. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Examination |
|
|
|
|
|
No formal interpretability or probing analysis was conducted. |
|
|
|
|
|
--- |
|
|
|
|
|
## Environmental Impact |
|
|
|
|
|
Carbon emissions were not formally measured. |
|
|
|
|
|
* **Hardware Type:** Consumer-grade GPU |
|
|
* **Hours used:** < 24 hours |
|
|
* **Cloud Provider:** None (local training) |
|
|
* **Compute Region:** Not applicable |
|
|
* **Carbon Emitted:** Not estimated |
|
|
|
|
|
--- |
|
|
|
|
|
## Technical Specifications |
|
|
|
|
|
### Model Architecture and Objective |
|
|
|
|
|
* Decoder-only Transformer |
|
|
* Next-token prediction objective |
|
|
* LoRA adapters applied to attention layers |
|
|
|
|
|
### Compute Infrastructure |
|
|
|
|
|
#### Hardware |
|
|
|
|
|
* Single consumer-grade GPU |
|
|
|
|
|
#### Software |
|
|
|
|
|
* PyTorch |
|
|
* Transformers |
|
|
* PEFT |
|
|
|
|
|
--- |
|
|
|
|
|
## Citation |
|
|
|
|
|
**BibTeX:** |
|
|
|
|
|
```bibtex |
|
|
@misc{darwinkernelpanic2025luau, |
|
|
title={CodeLlama 7B Instruct Luau LoRA}, |
|
|
author={darwinkernelpanic}, |
|
|
year={2025}, |
|
|
howpublished={Hugging Face}, |
|
|
note={LoRA fine-tuned for Luau / Roblox scripting} |
|
|
} |
|
|
``` |
|
|
|
|
|
**APA:** |
|
|
|
|
|
darwinkernelpanic. (2025). *CodeLlama 7B Instruct Luau LoRA*. Hugging Face. |
|
|
|
|
|
--- |
|
|
|
|
|
## Model Card Authors |
|
|
|
|
|
darwinkernelpanic |
|
|
|
|
|
## Model Card Contact |
|
|
|
|
|
Use the Hugging Face repository issues or the author’s profile. |
|
|
|
|
|
--- |
|
|
|
|
|
### Framework versions |
|
|
|
|
|
* PEFT 0.18.0 |