--- license: apache-2.0 task_categories: - question-answering - feature-extraction language: - en tags: - eda - analog - vlm pretty_name: Analog Layouts Dataset for Vision Language Models (VLMs) --- # A VLM Framework to Optimize the Analysis of Analog Circuit Layouts ***ICML 2026 Submission - Under Review*** This repository contains the dataset presented in the paper *"A VLM Framework to Optimize the Analysis of Analog Circuit Layouts"*, along with the code for training and evaluating Visual Language Models (VLMs) on analog circuit layouts analysis tasks. The project addresses the challenge of interpreting technical diagrams by benchmarking VLMs on tasks ranging from single device identification to component counting in complex mixed circuits. ## Dataset Overview The dataset comprises over **30,000 circuits** and **77,000+ Question-Answer pairs**, organized into a comprehensive benchmark suite. ### Circuit Categories - **Single Devices** (19,997 images): PMOS, NMOS, Capacitors, Resistors. - **Base Circuits** (5,894 images): Ahuja OTA, Gate Driver, HPF, LDO, LPF, Miller OTA. - **Mixed Circuits** (4,140 images): Complex combinations of base circuits. ### Benchmark Tasks The dataset defines 5 core tasks for evaluation: | Task | Description | Size | |------|-------------|------| | **Task A** | Single device identification | 19,997 samples | | **Task B** | Base circuit identification | 5,894 samples | | **Task C** | Component counting (base circuits) | 27,475 samples | | **Task D** | Component counting (mixed circuits) | 19,848 samples | | **Task E** | Base circuit identification in mixed circuits | 4,140 samples | ## Repository Structure Once `code.zip` and `dataset.zip` have been unzipped, the structure is as follows: ``` . ├── code/ # Source code for fine-tuning and inference ├── base_circuits/ # Base circuit datasets and templates ├── mixed_circuits/ # Mixed circuit datasets ├── single_devices/ # Single device datasets └── tasks/ # Task definitions and data splits ``` ## Getting Started ### Prerequisites All execution scripts are located in the `code/` directory. ```bash cd code pip install -r requirements.txt ``` ### Fine-Tuning The repository provides a sequential fine-tuning launcher to handle dataset ablations and multiple tasks. **Basic Usage:** ```bash # Dry-run to view planned training jobs python VLM_finetune/run_ablation_sequential_ft.py --dry_run # Train Task A (Single device identification) with 100% of dataset python VLM_finetune/run_ablation_sequential_ft.py --task a1 --perc 100 ``` **Advanced Usage:** Train multiple tasks with specific data percentages: ```bash python VLM_finetune/run_ablation_sequential_ft.py --tasks a1,b1,c1 --percs 25,50,75,100 ``` ### Evaluation The inference pipeline supports evaluating both base models and fine-tuned LoRA adapters. **Batch Evaluation (Ablation Study):** Evaluate many adapters across different tasks and splits: ```bash python VLM_inference/run_ft_eval_ablation.py \ --splits-root /path/to/dataset/ablation_splits \ --adapter-root /path/to/outputs/finetune_lora \ --cache-dir /path/to/cache ``` **Result Reorganization:** Map raw evaluation results from training tasks (A1/B1/C1) to the final benchmark tasks (A-E) and compute aggregated metrics: ```bash python reorganize_results.py \ --input-root /path/to/raw_results \ --output-root /path/to/final_results ``` **Single Task Evaluation:** Run inference on a single task/circuit: ```bash # Evaluate Task A (Task A1) python VLM_inference/test_base_models/run_ft_eval_update.py --task a1 --num-samples 200 # Evaluate with a specific adapter python VLM_inference/test_base_models/run_ft_eval_update.py \ --task a1 \ --num-samples 200 \ --adapter /path/to/adapter/checkpoint ```