Cloud Expert Qwen
This is a version 1 of Qwen 2.5-Coder 1.5B fine-tuned for cloud computing & IaC
Fine-tuned on comprehensive cloud computing, Infrastructure as Code, containerization, and Linux system administration documentation.
π― What This Model Knows
- Cloud Platforms: Azure, AWS, GCP
- Infrastructure as Code: Terraform, CloudFormation, ARM templates
- Containers & Orchestration: Docker, Kubernetes
- Linux: System administration, troubleshooting, networking
- DevOps: CI/CD, monitoring, security best practices
π Quick Start
Option 1: Use with Transformers (Python)
from transformers import AutoModelForCausalLM, AutoTokenizer
from peft import PeftModel
import torch
# Load base model
base_model = "Qwen/Qwen2.5-Coder-1.5B-Instruct"
model = AutoModelForCausalLM.from_pretrained(
base_model,
torch_dtype=torch.float16,
device_map="auto",
trust_remote_code=True
)
# Load fine-tuned LoRA adapters
model = PeftModel.from_pretrained(model, "jsdjsdequinia/cloud-expert-qwen/lora-adapters")
tokenizer = AutoTokenizer.from_pretrained(base_model, trust_remote_code=True)
# Ask a question
question = "How do I troubleshoot SSH connection issues on Linux?"
prompt = f"<|im_start|>user\n{question}<|im_end|>\n<|im_start|>assistant\n"
inputs = tokenizer(prompt, return_tensors="pt").to("cuda")
outputs = model.generate(**inputs, max_new_tokens=200, temperature=0.7)
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
print(response)
Option 2: Use with Ollama (Recommended for CPU)
Perfect for laptops without GPU!
Install Ollama: https://ollama.ai/download
Download the model:
huggingface-cli download jsdjsdequinia/cloud-expert-qwen cloud-expert-qwen-q8_0.gguf --local-dir ./
huggingface-cli download jsdjsdequinia/cloud-expert-qwen Modelfile --local-dir ./
- Create Ollama model:
ollama create cloud-expert -f Modelfile
- Run it:
ollama run cloud-expert
- Or use in code:
import ollama
response = ollama.chat(model='cloud-expert', messages=[
{'role': 'user', 'content': 'What is Azure Virtual Machine?'}
])
print(response['message']['content'])
π¦ Available Formats
| Format | Size | Use Case | Download |
|---|---|---|---|
| LoRA Adapters | ~100MB | Fine-tuning, GPU inference | lora-adapters/ |
| Merged Model | ~3GB | Full model, GPU inference | merged-model/ |
| GGUF (q8_0) | ~1.5GB | CPU inference with Ollama | *.gguf |
π Training Details
- Base Model: Qwen/Qwen2.5-Coder-1.5B-Instruct
- Training Method: LoRA (Low-Rank Adaptation)
- Training Data: 43 examples
- Manual Q&A pairs on cloud services
- Scraped official documentation (Azure, Docker, Kubernetes, etc.)
- Linux troubleshooting guides
- Training Time: ~20-30 minutes on RTX 3070
- Trainable Parameters: ~2% (LoRA efficient training)
π‘ Example Questions
- What is Microsoft Azure?
- How do I deploy a Docker container?
- Explain Terraform state management
- How do I troubleshoot disk usage on Linux?
- Compare Azure VM vs AWS EC2
- What are Kubernetes best practices?
- How do I configure a Linux firewall?
π₯οΈ System Requirements
For Training:
- GPU with 8GB+ VRAM
- 16GB RAM
- CUDA 12.1+
For Inference:
With Transformers (GPU):
- GPU with 4GB+ VRAM
- 8GB RAM
With Ollama (CPU - Recommended for work laptops):
- Any modern CPU
- 4GB RAM
- No GPU needed! β
β‘ Performance
| Setup | Tokens/Second | Use Case |
|---|---|---|
| GPU (RTX 3070) | ~50 tok/s | Development, training |
| CPU (Ollama, 16GB RAM) | ~10-15 tok/s | Work laptop, portable |
π Use Cases
β
Learning cloud technologies
β
Quick reference for DevOps tasks
β
Understanding IaC best practices
β
Linux troubleshooting assistance
β
Comparing cloud services
β
Interview preparation
β οΈ Limitations
- Knowledge cutoff: Training data as of 2024
- May not reflect very recent service updates
- Best for general concepts and established practices
- Always verify critical production decisions with official docs
- Not a replacement for hands-on experience
π License
Apache 2.0 - Free for commercial and personal use
π Credits
- Base model: Qwen/Qwen2.5-Coder-1.5B-Instruct
- Fine-tuned by: jsdjsdequinia
- Documentation sources: Microsoft Azure, Docker, Kubernetes, DigitalOcean, HashiCorp
π Feedback
Found an issue or have suggestions? Feel free to open an issue on the model page!
Built with β€οΈ for the DevOps community
- Downloads last month
- 35
8-bit
Model tree for jsdjsdequinia/cloud-expert-qwen
Base model
Qwen/Qwen2.5-1.5B