darwinkernelpanic's picture
Update README.md
489139c verified
---
base_model: codellama/CodeLlama-7b-Instruct-hf
library_name: peft
pipeline_tag: text-generation
tags:
- base_model:adapter:codellama/CodeLlama-7b-Instruct-hf
- lora
- transformers
- luau
- roblox
license: apache-2.0
language:
- en
---
# Model Card for CodeLlama-7B-Instruct-Luau
Fine-tuned version of `codellama/CodeLlama-7b-Instruct-hf` targeted toward the **Luau** programming language, Roblox’s Lua-derived scripting language.
This model is distributed as a **LoRA adapter** and is intended to improve the base model’s performance on Roblox-specific scripting tasks.
---
## Model Details
### Model Description
This model is a parameter-efficient fine-tuning (LoRA) of CodeLlama 7B Instruct, specialized for generating, explaining, and refactoring **Luau** code.
The fine-tuning focuses on Roblox development patterns, including common services, APIs, gameplay scripting idioms, and client/server logic. The model is designed to assist developers during prototyping, learning, and general scripting workflows.
- **Developed by:** darwinkernelpanic
- **Funded by:** Not applicable
- **Shared by:** darwinkernelpanic
- **Model type:** Causal Language Model (decoder-only, LoRA adapter)
- **Language(s) (NLP):** English
- **License:** Apache-2.0
- **Finetuned from model:** codellama/CodeLlama-7b-Instruct-hf
### Model Sources
- **Repository:** https://huggingface.co/darwinkernelpanic/CodeLlama-7b-Instruct-hf-luau
- **Paper:** *Code Llama: Large Language Models for Code* (Meta AI)
- **Demo:** Not available
---
## Uses
### Direct Use
This model can be used directly for:
- Writing Luau scripts for Roblox
- Explaining Roblox APIs and services
- Refactoring or debugging Luau code
- Prototyping gameplay systems and utilities
- Learning Luau and Roblox scripting concepts
The model is intended as a **developer assistant**, not an autonomous system.
### Downstream Use
Potential downstream uses include:
- Further fine-tuning on proprietary Roblox frameworks
- Integration into IDEs or editor tooling
- Chat-based assistants for Roblox development
- Educational or documentation tooling
### Out-of-Scope Use
This model should **not** be used for:
- Safety-critical or production-critical systems
- Legal, medical, or financial advice
- Malware, exploit, or cheat development
- Fully automated code deployment without review
---
## Bias, Risks, and Limitations
- Inherits biases and limitations from the base CodeLlama model
- May hallucinate Roblox APIs or outdated behaviors
- Does not validate code at runtime
- Output correctness depends on prompt quality
### Recommendations
Users should:
- Review all generated code manually
- Test scripts in Roblox Studio
- Cross-check with official Roblox documentation
- Treat outputs as suggestions rather than authoritative solutions
---
## How to Get Started with the Model
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
base_model = "codellama/CodeLlama-7b-Instruct-hf"
adapter_model = "darwinkernelpanic/CodeLlama-7b-Instruct-hf-luau"
tokenizer = AutoTokenizer.from_pretrained(base_model)
model = AutoModelForCausalLM.from_pretrained(base_model)
model = PeftModel.from_pretrained(model, adapter_model)
prompt = "Write a Luau function that creates a Part and parents it to Workspace."
inputs = tokenizer(prompt, return_tensors="pt")
outputs = model.generate(
**inputs,
max_new_tokens=300,
temperature=0.7,
do_sample=True
)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
````
---
## Training Details
### Training Data
The model was fine-tuned on a curated mixture of:
* Luau scripts
* Roblox API usage examples
* Open-source Roblox projects
* Synthetic instruction-style prompts
All data was filtered to avoid private, proprietary, or sensitive content.
### Training Procedure
The model was trained using parameter-efficient fine-tuning with LoRA while keeping the base model weights frozen.
#### Preprocessing
* Code formatting normalization
* Instruction-style prompt structuring
* Removal of low-quality or irrelevant samples
#### Training Hyperparameters
* **Training regime:** fp16 mixed precision
#### Speeds, Sizes, Times
* **Base model size:** ~7B parameters
* **Trainable parameters:** <1% (LoRA adapters only)
* **Adapter checkpoint size:** ~100–200 MB
---
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
* Hand-written Luau prompts
* Roblox-specific scripting scenarios
#### Factors
* Luau syntax correctness
* Roblox API familiarity
* Instruction-following behavior
#### Metrics
* Qualitative human evaluation
* Manual code review and comparison with base model
### Results
The LoRA adapter demonstrates improved performance over the base model in:
* Generating idiomatic Luau
* Correct Roblox service usage
* Following game-development-oriented instructions
#### Summary
The model performs best when used as a Roblox development assistant and is not intended for general-purpose natural language tasks.
---
## Model Examination
No formal interpretability or probing analysis was conducted.
---
## Environmental Impact
Carbon emissions were not formally measured.
* **Hardware Type:** Consumer-grade GPU
* **Hours used:** < 24 hours
* **Cloud Provider:** None (local training)
* **Compute Region:** Not applicable
* **Carbon Emitted:** Not estimated
---
## Technical Specifications
### Model Architecture and Objective
* Decoder-only Transformer
* Next-token prediction objective
* LoRA adapters applied to attention layers
### Compute Infrastructure
#### Hardware
* Single consumer-grade GPU
#### Software
* PyTorch
* Transformers
* PEFT
---
## Citation
**BibTeX:**
```bibtex
@misc{darwinkernelpanic2025luau,
title={CodeLlama 7B Instruct Luau LoRA},
author={darwinkernelpanic},
year={2025},
howpublished={Hugging Face},
note={LoRA fine-tuned for Luau / Roblox scripting}
}
```
**APA:**
darwinkernelpanic. (2025). *CodeLlama 7B Instruct Luau LoRA*. Hugging Face.
---
## Model Card Authors
darwinkernelpanic
## Model Card Contact
Use the Hugging Face repository issues or the author’s profile.
---
### Framework versions
* PEFT 0.18.0