llm-base / README.md
psssd-cat's picture
Add model card
5d78db1 verified
---
license: apache-2.0
base_model: Cagatayd/llama3.2-1B-Instruct-Egitim
tags:
- llama3
- fine-tuned
- docker
- lora
datasets:
- MattCoddity/dockerNLcommands
---
# Fine-tuned Llama 3.2 1B for Docker Commands
This model is a fine-tuned version of [Cagatayd/llama3.2-1B-Instruct-Egitim](https://huggingface.co/Cagatayd/llama3.2-1B-Instruct-Egitim) on the [MattCoddity/dockerNLcommands](https://huggingface.co/datasets/MattCoddity/dockerNLcommands) dataset.
## Training Details
- **Base Model:** Cagatayd/llama3.2-1B-Instruct-Egitim
- **Dataset:** MattCoddity/dockerNLcommands
- **Training Method:** LoRA (Low-Rank Adaptation)
- **Quantization:** 8-bit
- **LoRA Rank:** 32
- **Training Steps:** 60
- **Final Accuracy:** 80.69%
## Usage
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("psssd-cat/llm-base")
tokenizer = AutoTokenizer.from_pretrained("psssd-cat/llm-base")
# Example usage
text = "Find all running docker containers"
inputs = tokenizer(text, return_tensors="pt")
outputs = model.generate(**inputs, max_new_tokens=100)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
```
## Training Configuration
- Learning Rate: 3e-5
- Batch Size: 2 per device
- Gradient Accumulation Steps: 4
- FP16: Enabled
- Warmup Steps: 5
## Model Performance
- Evaluation Loss: 1.147
- Mean Token Accuracy: 80.69%
- Training Loss: 2.094
---
Generated with [Claude Code](https://claude.com/claude-code)