Liquid-WebSight / README.md
saadxsalman's picture
Update README.md
e9e4e57 verified
metadata
language:
  - en
license: apache-2.0
base_model: LiquidAI/LFM-2.5-450M-Instruct
pipeline_tag: image-text-to-text
tags:
  - vision-language
  - image-to-code
  - tailwind-css
  - web-development
  - unsloth
  - liquid-foundation-model
model_name: Liquid-Web (LFM2.5-VL-450M-WebSight)
datasets:
  - HuggingFaceM4/WebSight

Model Card: Liquid-Web (LFM2.5-VL-450M-WebSight)

Model Details

  • Model Type: Vision-Language Model (VLM)
  • Base Model: LiquidAI/LFM2.5-VL-450M
  • Architecture: Liquid Foundation Model (LFM) with SigLIP2 Vision Encoder.
  • Fine-tuning Method: LoRA (Low-Rank Adaptation)
  • Task: Image-to-HTML/Tailwind CSS Generation (UI-to-Code)
  • Language: English
  • License: Apache 2.0 (Inherited from base)

Intended Use

Primary Use Case

This model is designed to take screenshots of web pages as input and generate the corresponding HTML code using Tailwind CSS utilities. It is intended for developers looking to automate the conversion of UI designs or existing web pages into clean, functional code.

Out-of-Scope Use

  • General-purpose image captioning (it is highly specialized for code).
  • Generating scripts for malicious web automation.
  • Production-ready code without human review (the model may hallucinate specific color shades or precise pixel alignments).

Training Data

The model was fine-tuned on the HuggingFaceM4/WebSight (v0.2) dataset.

  • Format: Paired images (screenshots) and text (HTML + Tailwind CSS).
  • Volume: 1,000 high-quality samples (fine-tuned subset).
  • Content: Diverse layouts including landing pages, dashboards, and portfolio sites.

Training Procedure

Hyperparameters

Parameter Value
LoRA Rank (r) 64
LoRA Alpha 64
Optimizer AdamW (8-bit)
Learning Rate 2e-4
Batch Size 1 (Per device)
Gradient Accumulation 8
Max Steps 100
Precision 4-bit Quantization (NormalFloat4)

Frameworks Used:

  • Unsloth: For memory-efficient training and fast kernels.
  • TRL & PEFT: For the Supervised Fine-Tuning (SFT) loop.

Performance & Limitations

Strengths

  • Extreme Efficiency: At only 450M parameters, this model provides high-speed inference on edge devices (Jetson Orin, Mobile).
  • Modern CSS: Specifically trained on Tailwind CSS v2.2+, avoiding bloated traditional CSS files.

Limitations

  • Hallucinations: The model may occasionally invent Tailwind classes that do not exist (e.g., text-custom-500).
  • Complexity: Very deep or complex nested layouts might result in truncated HTML due to the max_seq_length limit.
  • Resolution: Fine details in very small text within screenshots may be missed by the vision encoder.

How to Get Started

To use this model for inference, use the following code snippet:

from unsloth import FastLanguageModel
from transformers import AutoProcessor
from PIL import Image

model, tokenizer = FastLanguageModel.from_pretrained("saadxsalman/LFM-WebSight-Tailwind", load_in_4bit=True)
processor = AutoProcessor.from_pretrained("saadxsalman/LFM-WebSight-Tailwind")

# Inference logic goes here

Citation

If you use this model, please cite the original WebSight technical report:

@misc{laurençon2024unlocking,
      title={Unlocking the conversion of Web Screenshots into HTML Code with the WebSight Dataset}, 
      author={Hugo Laurençon and Léo Tronchon and Victor Sanh},
      year={2024},
      eprint={2403.09029},
      archivePrefix={arXiv},
      primaryClass={cs.HC}
}