--- license: apache-2.0 base_model: Cagatayd/llama3.2-1B-Instruct-Egitim tags: - llama3 - fine-tuned - docker - lora datasets: - MattCoddity/dockerNLcommands --- # Fine-tuned Llama 3.2 1B for Docker Commands This model is a fine-tuned version of [Cagatayd/llama3.2-1B-Instruct-Egitim](https://huggingface.co/Cagatayd/llama3.2-1B-Instruct-Egitim) on the [MattCoddity/dockerNLcommands](https://huggingface.co/datasets/MattCoddity/dockerNLcommands) dataset. ## Training Details - **Base Model:** Cagatayd/llama3.2-1B-Instruct-Egitim - **Dataset:** MattCoddity/dockerNLcommands - **Training Method:** LoRA (Low-Rank Adaptation) - **Quantization:** 8-bit - **LoRA Rank:** 32 - **Training Steps:** 60 - **Final Accuracy:** 80.69% ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("psssd-cat/llm-base") tokenizer = AutoTokenizer.from_pretrained("psssd-cat/llm-base") # Example usage text = "Find all running docker containers" inputs = tokenizer(text, return_tensors="pt") outputs = model.generate(**inputs, max_new_tokens=100) print(tokenizer.decode(outputs[0], skip_special_tokens=True)) ``` ## Training Configuration - Learning Rate: 3e-5 - Batch Size: 2 per device - Gradient Accumulation Steps: 4 - FP16: Enabled - Warmup Steps: 5 ## Model Performance - Evaluation Loss: 1.147 - Mean Token Accuracy: 80.69% - Training Loss: 2.094 --- Generated with [Claude Code](https://claude.com/claude-code)