--- license: apache-2.0 --- # Vision-Language Pairs Dataset This dataset contains metadata about image-text pairs from various popular vision-language datasets. ## Contents - **vision_language_data/all_vision_language_images.csv**: Combined metadata for all images (75629 records) - **vision_language_data/all_vision_language_captions.csv**: Combined captions for all images (86676 records) - **dataset_statistics.csv**: Summary statistics for each dataset - **category_distribution.csv**: Distribution of image categories across datasets - **caption_length_distribution.csv**: Distribution of caption lengths - **caption_style_distribution.csv**: Distribution of caption styles - **category_caption_statistics.csv**: Caption statistics by category - **vision_language_catalog.json**: Searchable catalog with sample image-caption pairs ## Datasets Included - **COCO** (Common Objects in Context): COCO is a large-scale object detection, segmentation, and captioning dataset with multiple captions per image. (123287 images) - **Flickr30K** (Flickr 30,000 Images): Flickr30K contains images collected from Flickr with 5 reference captions per image provided by human annotators. (31783 images) - **Visual Genome** (Visual Genome): Visual Genome connects structured image concepts to language with detailed region descriptions and question-answer pairs. (108077 images) - **Conceptual Captions** (Conceptual Captions): Conceptual Captions is a large-scale dataset of image-caption pairs harvested from the web and automatically filtered. (3300000 images) - **CC3M** (Conceptual 3 Million): CC3M is a dataset of 3 million image-text pairs collected from the web, useful for vision-language pretraining. (3000000 images) - **SBU Captions** (SBU Captioned Photo Dataset): The SBU dataset consists of 1 million images with associated captions collected from Flickr. (1000000 images) ## Fields Description ### Images Table - **image_id**: Unique identifier for the image - **dataset**: Source dataset name - **image_url**: URL to the image (simulated) - **primary_category**: Main content category - **width**: Image width in pixels - **height**: Image height in pixels - **aspect_ratio**: Width divided by height - **caption_count**: Number of captions for this image - **license**: License under which the image is available ### Captions Table - **caption_id**: Unique identifier for the caption - **image_id**: ID of the associated image - **dataset**: Source dataset name - **text**: Caption text - **language**: Caption language (default: en) - **style**: Caption style (descriptive, short, or detailed) - **length**: Number of characters in the caption - **word_count**: Number of words in the caption ## Usage Examples This metadata can be used for: 1. Analyzing the composition of vision-language datasets 2. Comparing caption characteristics across different datasets 3. Training and evaluating image captioning models 4. Studying linguistic patterns in image descriptions 5. Developing multimodal AI systems ## Data Generation Note This dataset contains synthetic metadata that represents the structure and characteristics of actual vision-language pair collections, but the specific image and caption details are generated for demonstration purposes. Created: 2025-04-26 ## Note: ## All files are packaged into a ZIP archive called vision_language_pairs_data.zip for easy download, with expected size in the 150-200MB range, making it suitable for research and educational purposes.