--- language: - en license: apache-2.0 base_model: Qwen/Qwen3-0.6B tags: - qwen3 - fine-tuned - web-development - coding - sft pipeline_tag: text-generation --- # qwen3-webdev-0.6b A fine-tuned version of [Qwen/Qwen3-0.6B](https://huggingface.co/Qwen/Qwen3-0.6B) on a curated dataset of real-world web development Q&A. ## Model Description This model is fine-tuned to answer junior-to-mid-level web development questions covering HTML, CSS, JavaScript, React, APIs, and common frontend/backend concepts. - **Base model:** Qwen/Qwen3-0.6B - **Fine-tuning method:** Supervised Fine-Tuning (SFT) with TRL - **Dataset:** 307 real web development Q&A pairs (interview-style) - **Training:** 3 epochs, final loss 0.7072 - **Hardware:** NVIDIA RTX 4090 Mobile (16GB) ## Intended Use - Learning tool for web development concepts - Junior dev quick-reference assistant - Demo of efficient small-model fine-tuning pipeline ## Training Details | Parameter | Value | |---|---| | Base model | Qwen3-0.6B | | Dataset size | 307 examples | | Epochs | 3 | | Final train loss | 0.7072 | | Precision | bfloat16 | ## Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch model = AutoModelForCausalLM.from_pretrained("PacificDev/qwen3-webdev-0.6b", dtype=torch.bfloat16) tokenizer = AutoTokenizer.from_pretrained("PacificDev/qwen3-webdev-0.6b") prompt = "What is the difference between flexbox and CSS grid?" inputs = tokenizer(f"Question: {prompt}\nAnswer:", return_tensors="pt") output = model.generate(**inputs, max_new_tokens=300, temperature=0.7, do_sample=True) print(tokenizer.decode(output[0], skip_special_tokens=True)) ``` ## Limitations - Small model (0.6B params) — answers are concise/simplified - Dataset is limited to 307 examples — may not cover all topics - Outputs `` reasoning tags (Qwen3 chain-of-thought) - Not suitable for production use without further evaluation ## License Apache 2.0 (same as base model)