JessicaE's picture
Update README.md
39c1d3a verified
---
dataset_info:
features:
- name: file_name
dtype: string
- name: source_file
dtype: string
- name: question
dtype: string
- name: question_type
dtype: string
- name: question_id
dtype: int32
- name: answer
dtype: string
- name: answer_choices
list: string
- name: correct_choice_idx
dtype: int32
- name: image
dtype: image
- name: video
dtype: video
- name: media_type
dtype: string
splits:
- name: test
num_bytes: 187015546578
num_examples: 102678
download_size: 175022245655
dataset_size: 187015546578
configs:
- config_name: default
data_files:
- split: test
path: data/test-*
license: mit
task_categories:
- visual-question-answering
language:
- en
size_categories:
- 100K<n<1M
---
# OpenSeeSimE-Structural: Engineering Simulation Visual Question Answering Benchmark
## Dataset Summary
OpenSeeSimE-Structural is a large-scale benchmark dataset for evaluating vision-language models on structural analysis simulation interpretation tasks. It contains over 100,000 question-answer pairs across parametrically-varied structural simulations including stress analysis, and deformation patterns.
## Purpose
While vision-language models (VLMs) have shown promise in general visual reasoning, their effectiveness for specialized engineering simulation interpretation remains largely unexplored. This benchmark enables:
- Statistically robust evaluation of VLM performance on engineering visualizations
- Assessment across multiple reasoning capabilities (captioning, reasoning, grounding, relationship understanding)
- Evaluation using different question types (binary classification, multiple-choice, spatial grounding)
## Dataset Composition
### Statistics
- **Total instances**: 102,678 question-answer pairs
- **Simulation types**: 5 structural models (Dog Bone, Hip Implant, Pressure Vessel, Beams, Wall Bracket)
- **Parametric variations**: 1,024 unique instances per base model (4^5 parameter combinations)
- **Question categories**: Captioning, Reasoning, Grounding, Relationship Understanding
- **Question types**: Binary, Multiple-choice, Spatial grounding
- **Media formats**: Both static images (1920×1440 PNG) and videos (Originally Extracted at: 200 frames, 29 fps, 7 seconds)
### Simulation Parameters
Each base model varies across 5 parameters with 4 values each:
**Dog Bone**: Length, Thickness, Diameter, Axial Load, Bending Load
**Hip Implant**: Beam Length, Beam Diameter, Ball Diameter, Axial Load, Bending Load
**Pressure Vessel**: Length, Thickness, Diameter, Material, Pressure
**Thermal Beam**: Thickness, Bending Load, Young's Modulus, Tensile Yield Strength, Cross Section Shape
**Wall Bracket**: Length, Width, Height, Thickness, Bending Force
### Question Distribution
- **Binary Classification**: 40% (yes/no questions about symmetry, stress types, uniformity, etc.)
- **Multiple-Choice**: 30% (4-option questions about deformation direction, stress dominance, magnitude ranges, etc.)
- **Spatial Grounding**: 30% (location-based questions with labeled regions A/B/C/D)
## Data Collection Process
### Simulation Generation
1. Base models sourced from Ansys Mechanical tutorial files
2. Parametric automation via PyMechanical and PyGeometry interfaces
3. Systematic variation across 5 parameters with 4 linearly-spaced values
4. All simulations solved using finite element analysis with validated convergence settings
### Ground Truth Extraction
Automated extraction eliminates human annotation costs and ensures consistency:
- **Statistical Analysis**: Direct queries on result arrays (max, min, mean, std)
- **Distribution Analysis**: Threshold-based classification using coefficient of variation
- **Physics-Based Classification**: Stress tensor analysis and mechanics principles
- **Spatial Localization**: Color-based region generation with computer vision algorithms
All ground truth derived from numerical simulation results rather than visual interpretation.
## Preprocessing and Data Format
### Image Processing
- Resolution: 1920×1440 pixels
- Format: PNG with lossless compression
- Standardized viewing orientations: front, back, left, right, top, bottom, isometric
- Consistent color mapping: rainbow gradients (red=maximum, blue=minimum)
- Automatic deformation scaling (1.5× relative to maximum dimension)
### Video Processing
- 200 frames at 29 fps (7 seconds duration)
- Maximum deformation at frame 100 (temporal midpoint)
- H.264 compression at 1920×1440 resolution
- Uniform frame sampling for model input (32 frames)
### Data Fields
```python
{
'file_name': str, # Unique identifier
'source_file': str, # Base simulation model
'question': str, # Question text
'question_type': str, # 'Binary', 'Multiple Choice', or 'Spatial'
'question_id': int, # Question identifier (1-20)
'answer': str, # Ground truth answer
'answer_choices': List[str], # Available options
'correct_choice_idx': int, # Index of correct answer
'image': Image, # PIL Image object (1920×1440)
'video': Video, # Video frames
'media_type': str # 'image' or 'video'
}
```
## Labels
All labels are automatically generated from simulation numerical results:
- **Binary questions**: "Yes" or "No"
- **Multiple-choice**: Single letter (A/B/C/D) or descriptive option
- **Spatial grounding**: Region label (A/B/C/D) corresponding to labeled visualization locations
Label generation employs domain-specific thresholds:
- Uniformity: CV ≤ 0.2 (20%)
- Symmetry: 60% of node pairs within 10% tolerance (structural)
- Spatial matching: 50-pixel separation for region placement
## Dataset Splits
- **Test split only**: 102,678 instances
- No train/validation splits provided (evaluation benchmark, not for model training)
- Representative sampling across all simulation types and question categories
## Intended Use
### Primary Use Cases
1. **Benchmark evaluation** of vision-language models on engineering simulation interpretation
2. **Capability assessment** across visual reasoning dimensions (captioning, spatial grounding, relationship understanding)
3. **Transfer learning analysis** from general-domain to specialized technical visual reasoning
### Out-of-Scope Use
- Real-time engineering decision-making without expert validation
- Safety-critical applications without human oversight
- Generalization to simulation types beyond structural mechanics
## Limitations
### Technical Limitations
- **Objective tasks only**: Excludes subjective engineering judgments requiring domain expertise
- **Single physics domain**: Structural mechanics only (see OpenSeeSimE-Fluid for fluid dynamics)
- **Ansys-specific**: Visualizations generated using Ansys Mechanical rendering conventions
- **Static parameters**: Fixed material properties and boundary conditions per instance
- **2D visualizations**: All inputs are 2D projections of 3D simulations
### Known Biases
- **Color scheme dependency**: Questions exploit default rainbow gradient conventions
- **Geometry bias**: Selected simulation types may not represent full diversity of structural analysis applications
- **View orientation bias**: Standardized camera positions may not capture all critical simulation features
## Ethical Considerations
### Responsible Use
- Models evaluated on this benchmark should NOT be deployed for safety-critical engineering decisions without expert validation
- Automated interpretation should augment, not replace, human engineering expertise
- Users should verify that benchmark performance translates to their specific simulation contexts
### Data Privacy
- All simulations contain no proprietary or confidential engineering data
- No personal information collected
- Publicly available tutorial files used as base models
### Environmental Impact
- Dataset generation required significant computational CPU resources
- Consider environmental cost of large-scale model evaluation on this benchmark
## License
MIT License - Free for academic and commercial use with attribution
## Citation
If you use this dataset, please cite:
```bibtex
@article{ezemba2024opensesime,
title={OpenSeeSimE: A Large-Scale Benchmark to Assess Vision-Language Model Question Answering Capabilities in Engineering Simulations},
author={Ezemba, Jessica and Pohl, Jason and Tucker, Conrad and McComb, Christopher},
year={2025}
}
```
## AI Usage Disclosure
### Dataset Generation
- **Simulation automation**: Python scripts with Ansys PyMechanical interface
- **Ground truth extraction**: Automated computational protocols (no AI involvement)
- **Quality validation**: Expert oversight of automated extraction procedures
- **No generative AI** used in dataset creation, labeling, or curation
### Visualization Generation
- Ansys Mechanical rendering engine (deterministic, physics-based)
- Standardized color mapping and camera controls
- No AI-based image generation or enhancement
## Contact
**Authors**: Jessica Ezemba (jezemba@andrew.cmu.edu), Jason Pohl, Conrad Tucker, Christopher McComb
**Institution**: Department of Mechanical Engineering, Carnegie Mellon University
## Acknowledgments
- Ansys for providing simulation software and tutorial files
- Carnegie Mellon University for computational resources
- Reviewers and domain experts who validated the automated extraction protocols
---
**Version**: 1.0
**Last Updated**: December 2025
**Status**: Complete and stable