| --- |
| license: apache-2.0 |
| task_categories: |
| - question-answering |
| - feature-extraction |
| language: |
| - en |
| tags: |
| - eda |
| - analog |
| - vlm |
| pretty_name: Analog Layouts Dataset for Vision Language Models (VLMs) |
| --- |
| |
| # THEIA: A Multimodal Dataset and Benchmark for Vision-Language Analysis of Layout |
|
|
| ***NeurIPS 2026 Evaluations and Datasets Submission - Under Review*** |
|
|
| This repository contains the dataset presented in the paper *"THEIA: A Multimodal Dataset and Benchmark for Vision-Language Analysis of Layout"*, along with the code for training and evaluating Visual Language Models (VLMs) on analog circuit layouts analysis tasks. |
|
|
| The project addresses the challenge of interpreting technical diagrams by benchmarking VLMs on tasks ranging from single device identification to component counting in complex mixed circuits. |
|
|
| ## Dataset Overview |
|
|
| The dataset comprises over **30,000 circuits** and **77,000+ Question-Answer pairs**, organized into a comprehensive benchmark suite. |
|
|
| ### Circuit Categories |
| - **Single Devices** (19,997 images): PMOS, NMOS, Capacitors, Resistors. |
| - **Base Circuits** (5,894 images): Ahuja OTA, Gate Driver, HPF, LDO, LPF, Miller OTA. |
| - **Mixed Circuits** (4,140 images): Complex combinations of base circuits. |
|
|
| ### Benchmark Tasks |
| The dataset defines 5 core tasks for evaluation: |
| | Task | Description | Size | |
| |------|-------------|------| |
| | **Task A** | Single device identification | 19,997 samples | |
| | **Task B** | Base circuit identification | 5,894 samples | |
| | **Task C** | Component counting (base circuits) | 27,475 samples | |
| | **Task D** | Component counting (mixed circuits) | 19,848 samples | |
| | **Task E** | Base circuit identification in mixed circuits | 4,140 samples | |
|
|
| ## Repository Structure |
| Once `code.zip` and `dataset.zip` have been unzipped, the structure is as follows: |
| ``` |
| . |
| ├── code/ # Source code for fine-tuning and inference |
| ├── base_circuits/ # Base circuit datasets and templates |
| ├── mixed_circuits/ # Mixed circuit datasets |
| ├── single_devices/ # Single device datasets |
| └── tasks/ # Task definitions and data splits |
| ``` |
|
|
| ## Getting Started |
|
|
| ### Prerequisites |
| All execution scripts are located in the `code/` directory. |
|
|
| ```bash |
| cd code |
| pip install -r requirements.txt |
| ``` |
|
|
| ### Fine-Tuning |
| The repository provides a sequential fine-tuning launcher to handle dataset ablations and multiple tasks. |
|
|
| **Basic Usage:** |
| ```bash |
| # Dry-run to view planned training jobs |
| python VLM_finetune/run_ablation_sequential_ft.py --dry_run |
| |
| # Train Task A (Single device identification) with 100% of dataset |
| python VLM_finetune/run_ablation_sequential_ft.py --task a1 --perc 100 |
| ``` |
|
|
| **Advanced Usage:** |
| Train multiple tasks with specific data percentages: |
| ```bash |
| python VLM_finetune/run_ablation_sequential_ft.py --tasks a1,b1,c1 --percs 25,50,75,100 |
| ``` |
|
|
| ### Evaluation |
| The inference pipeline supports evaluating both base models and fine-tuned LoRA adapters. |
|
|
| **Batch Evaluation (Ablation Study):** |
| Evaluate many adapters across different tasks and splits: |
| ```bash |
| python VLM_inference/run_ft_eval_ablation.py \ |
| --splits-root /path/to/dataset/ablation_splits \ |
| --adapter-root /path/to/outputs/finetune_lora \ |
| --cache-dir /path/to/cache |
| ``` |
|
|
| **Result Reorganization:** |
| Map raw evaluation results from training tasks (A1/B1/C1) to the final benchmark tasks (A-E) and compute aggregated metrics: |
| ```bash |
| python reorganize_results.py \ |
| --input-root /path/to/raw_results \ |
| --output-root /path/to/final_results |
| ``` |
|
|
| **Single Task Evaluation:** |
| Run inference on a single task/circuit: |
| ```bash |
| # Evaluate Task A (Task A1) |
| python VLM_inference/test_base_models/run_ft_eval_update.py --task a1 --num-samples 200 |
| |
| # Evaluate with a specific adapter |
| python VLM_inference/test_base_models/run_ft_eval_update.py \ |
| --task a1 \ |
| --num-samples 200 \ |
| --adapter /path/to/adapter/checkpoint |
| ``` |