|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: qwen3-0.6B |
|
|
tags: |
|
|
- code-generation |
|
|
- svg |
|
|
- fine-tuned |
|
|
- fp16 |
|
|
- vllm |
|
|
- merged |
|
|
language: |
|
|
- en |
|
|
pipeline_tag: text-generation |
|
|
library_name: transformers |
|
|
model_type: qwen |
|
|
inference: true |
|
|
torch_dtype: float16 |
|
|
widget: |
|
|
- example_title: "Simple Circle" |
|
|
text: "Create a red circle" |
|
|
- example_title: "Rectangle with Border" |
|
|
text: "Draw a blue rectangle with black border" |
|
|
- example_title: "Complex Shape" |
|
|
text: "Generate a star with 5 points in yellow" |
|
|
--- |
|
|
|
|
|
# SVG Code Generator |
|
|
|
|
|
This is a fine-tuned model for generating SVG code from natural language descriptions. The model has been merged with the base model weights and optimized in fp16 format. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
- **Model Name**: model_v15 |
|
|
- **Base Model**: qwen3-0.6B |
|
|
- **Training Method**: Fine-tuning with merged weights |
|
|
- **Task**: Text-to-SVG code generation |
|
|
- **Model Type**: Merged Qwen model |
|
|
- **Precision**: fp16 |
|
|
- **Library**: Transformers, vLLM compatible |
|
|
- **Format**: Merged model (not adapter-based) |
|
|
|
|
|
## Usage |
|
|
|
|
|
### With Transformers |
|
|
|
|
|
Load the model directly using the transformers library: |
|
|
|
|
|
```python |
|
|
# Load base model and tokenizer |
|
|
from transformers import AutoTokenizer, AutoModelForCausalLM |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained("vinoku89/svg-code-generator") |
|
|
model = AutoModelForCausalLM.from_pretrained("vinoku89/svg-code-generator") |
|
|
|
|
|
|
|
|
# Generate SVG code |
|
|
prompt = "Create a blue circle with radius 50" |
|
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
|
|
|
|
# Generate with parameters |
|
|
outputs = model.generate( |
|
|
**inputs, |
|
|
max_length=200, |
|
|
temperature=0.7, |
|
|
do_sample=True, |
|
|
pad_token_id=tokenizer.eos_token_id |
|
|
) |
|
|
|
|
|
# Decode the generated SVG code |
|
|
generated_text = tokenizer.decode(outputs[0], skip_special_tokens=True) |
|
|
svg_code = generated_text[len(prompt):].strip() |
|
|
|
|
|
print("Generated SVG:") |
|
|
print(svg_code) |
|
|
``` |
|
|
|
|
|
### With vLLM |
|
|
|
|
|
This model supports vLLM for high-performance inference in fp16 format. |
|
|
|
|
|
## Training Data |
|
|
|
|
|
The model was trained on SVG code generation tasks with natural language descriptions. |
|
|
|
|
|
## Intended Use |
|
|
|
|
|
This model is designed to generate SVG code from text descriptions for educational and creative purposes. |
|
|
|
|
|
## Limitations |
|
|
|
|
|
- Generated SVG may require validation |
|
|
- Performance depends on prompt clarity |
|
|
- Limited to SVG syntax and features seen during training |
|
|
|
|
|
## Model Performance |
|
|
|
|
|
The model has been fine-tuned specifically for SVG generation tasks with merged weights for optimal performance. |
|
|
|
|
|
## Technical Details |
|
|
|
|
|
- **Precision**: fp16 for memory efficiency |
|
|
- **Compatibility**: vLLM supported for high-throughput inference |
|
|
- **Architecture**: Merged fine-tuned weights (no adapters required) |
|
|
|